Literature DB >> 25436085

Using eye movement analysis to study auditory effects on visual memory recall.

Ramtin Zargari Marandi1, Seyed Hojjat Sabzpoushan1.   

Abstract

Recent studies in affective computing are focused on sensing human cognitive context using biosignals. In this study, electrooculography (EOG) was utilized to investigate memory recall accessibility via eye movement patterns. 12 subjects were participated in our experiment wherein pictures from four categories were presented. Each category contained nine pictures of which three were presented twice and the rest were presented once only. Each picture presentation took five seconds with an adjoining three seconds interval. Similarly, this task was performed with new pictures together with related sounds. The task was free viewing and participants were not informed about the task's purpose. Using pattern recognition techniques, participants' EOG signals in response to repeated and non-repeated pictures were classified for with and without sound stages. The method was validated with eight different participants. Recognition rate in "with sound" stage was significantly reduced as compared with "without sound" stage. The result demonstrated that the familiarity of visual-auditory stimuli can be detected from EOG signals and the auditory input potentially improves the visual recall process.

Entities:  

Keywords:  Auditory effect; Cognitive Context; Electrooculography (EOG); Eye Movement Analysis; Memory Recall

Year:  2014        PMID: 25436085      PMCID: PMC4202595     

Source DB:  PubMed          Journal:  Basic Clin Neurosci        ISSN: 2008-126X


1. Introduction

Considerable advances in sensing, inferring, and using context information were achieved by investigating different dimensions of context, such as physical activity (Davies et al., 2008), location (Want et al., 1992), or the psychophysiological and affective state of studied subjects (Healey et al., 2010). Cognitive context of a person lies beyond these common contextual dimensions not necessarily providing a complete background context of a person. Based on the insights from the experimental psychology, the cognitive context comprises almost all aspects of mental processing, such as perception, memory, knowledge, and learning (Bulling & Roggen, 2011). Recent context-aware systems have a long way to cognitive context assessment in an unobtrusive manner. This is due to the fact that the cognitive context is encoded in complex neural dynamics inside the brain and few obvious cues are accessible by non-invasive measurement techniques (Bulling & Roggen, 2011). Cognitive neuroscience uses techniques such as functional magnetic resonance imaging (fMRI), (Chadwick et al., 2010) that are not suited for real-world applications. More potentially useful techniques investigating the cognitive context, such as electroencephalography EEG, (Bigdely-Shamlo et al., 2008), are not unobtrusive and robust enough for daily life setting applications. A large body of research in experimental psychology has shown that, in addition to physical activity, visual behavior is tightly linked with cognitive processes, such as attention (Liversedge & Findlay, 2000), memory (Hannula & Ranganath, 2009), learning (Heisz & Shore, 2008), or saliency determination (Henderson, 2003). Moreover, a firm relationship between eye movements and cognition makes eye movements a particularly promising source of information on the cognitive context of a person, beyond the mere physical or visual activities. A scenario to gain a good perspective on eye-based cognition-awareness was given by Bulling & Roggen, 2011, where attendees of a business reception wear eye trackers that are unobtrusively embedded into their goggles. By analyzing their eye movement patterns during conversations, cognition-aware memory assistants running on their mobile phones assess whether the involved speakers have met before and still remember each other. Using this information, the systems then automatically provided real-time memory assistance about people fallen into oblivion to prevent the embarrassing situations. Eye movement research is of great interest in neuroscience and psychiatry studies, as well as ergonomics, advertising and design as a window into observers’ visual and cognitive processes. For instance, researchers have utilized eye tracking to study behavior in such domains as image scanning e.g. (Noton & Stark, 1971), driving (Land & Lee, 1994), arithmetic (Suppes, 1990), analogy (Salvucci & Anderson, 2001), and reading (Rayner, 1998). In these domains as well as others, researchers typically analyze eye movements in terms of fixations (pauses over informative regions of interest) and saccades (rapid movements between fixations). Many researchers have started to study eye movements in natural environments in order to better understand the role which the visual system plays in the execution of everyday tasks (Hayhoe & Ballard, 2005). Human vision research has shown that unconscious eye movements are strongly related to the underlying cognitive and perceptive processes. For example, it has been shown that visual behavior is a good measure of visual engagement (Skotte et al., 2007), drowsiness (Schleicher et al. 2008), and cognitive load (Stuyven et al., 2000). Heisz et al. investigated changes in eye movement behavior across several exposures to face images (Heisz & Shore, 2008). They found that once a face becames more familiar, observers look longer and more often at the eyes and less often at the nose, mouth or forehead. Furthermore, it has been demonstrated that differences in eye movement patterns are linked to a number of mental disorders. Given this, eye tracking has been used to diagnose autism spectrum disorders (Boraston & Blakemore, 2007). For instance, Klin et al. showed that people with autism tend to show fewer fixations to the eyes but more to the mouth (Klin et al., 2002). Similar links were found for schizophrenia (Ettinger et al., 2006) as well as Parkinson's (Mosimann et al. 2005) and Alzheimer's disease (Crawford et al., 2005). All these studies suggest a close relationship between the visual behavior and cognition. These findings underline the potential of eye movement analysis for cognitive context assessment. However no attempts were done to detect familiarity from the eye movements whether for example a particular face, was previously seen and remembered. Along these lines, the emerging novel technologies which enable research on human behavior during complex cognitive processes like memory recall is drawing much more attentions nowadays.

2. Methods

An experimental paradigm was used to the study decision making process. the experiment conducted during this study comprised visual stimuli presentation to participants while recording their EOG signals. As it shown in Fig. 1, the experiment was conducted in a dim-lit, quiet room. For distinct and near to real world eye movements, visual stimuli were presented on a 116×66 cm screen by a video projector. The resolution of presented pictures as visual stimuli was 640×480 pixels. The stimuli were designed and implemented using Adobe Flash software. Participants were instructed on how to do the tasks using the interactive multimedia instructions running on a computer before the experiment. Moreover, instructions were also shown to the participants in text format before the task onset. Participants were requested to sit on an armchair and adjust their position such that their eyes were facing the center of the screen. Participant's distance to the screen was between 160 and 170 cm. The experiment was conducted in Brain-Computer Interface laboratory of Iran Neural Technology Research Centre.
Figure 1

A participant during experiment

A participant during experiment Eight picture sets were created, two sets for each of the following picture categories: abstract, landscapes, faces, and buildings images. Face pictures were selected from the IMM Face Database (Stegmann et al., 2003). Pictures with frontal view and neutral expression with dark background were chosen for face category; while pictures of the other categories were randomly selected from the Internet. However, it was ensured that these pictures had similar visual features. For example, landscape photographs that showed a lake as their main feature were selected and the building photographs always showed skyscrapers centered in the picture. Four sound sets were chosen and presented in association with the four picture categories. The sounds were chosen so that to help the memory recall process. Sounds of car passing, seashore, crowd talking and radio messages of spacemen were selected for building, landscape, face and abstract pictures, respectively.

2.1. Experiment

The experiment was performed in two sections. In the first section participants were presented four categories of pictures including faces, landscapes, abstract images and buildings without sound (Fig. 2). Each category contained nine pictures. Six pictures of each category were presented once and the remaining three pictures were presented twice. The pictures were presented randomly. As shown in Fig. 2, each presentation took five seconds after which a black screen with a dot-shaped down- counter at the center was presented for three seconds. The second section of the experiment immediately followed the first part. The second part was similar to the first one while the pictures were different and soundmatched. Participant's task in this experiment was free viewing. They were not informed about the aim of the experiment.
Figure 2

Sample pictures of the four categories (abstract, landscape, face, and building) and their sequence of display used in the experiment. Each picture was shown for five seconds; black screen pictures were shown in between for three seconds as intervals.

Sample pictures of the four categories (abstract, landscape, face, and building) and their sequence of display used in the experiment. Each picture was shown for five seconds; black screen pictures were shown in between for three seconds as intervals.

2.2. Participants

Fourteen participants (6 females and 8 males) volunteered for this experiment. Two male participants were excluded due to low quality EOG signal that prevented robust detection of eye movements. The remaining participants were 22 to 29 years old (M=25.8, SD=±3.6). All participants had normal or corrected-to-normal vision. Most of the participants were either undergraduate or graduate students from Iran University of Science and Technology.

2.3. Apparatus

Neurophysiologic signals were recorded and filtered using a Simulink application. For the present study, a g.USBAmp amplifier (g.tec Medical Engineering GmbH, Austria) was used. This device represents a multimodal amplifier for electrophysiological signals such as EEG, EOG, EMG and ECG. EOG signals were captured by 5 electrodes placed as in Fig. 3 and were recorded using a bipolar configuration. For the horizontal EOG (HEOG) the electrodes were placed at the outer canthus of each eye and for the vertical EOG (VEOG) infra-orbital and supra-orbital electrodes were placed in line with the pupil of one eye. Ground electrode was placed on the earlobe. Sampling rate was adjusted to 256 Hz. High-purity gold (Au) electrodes with resistance of lower than 5 kΩ were used in this research.
Figure 3

Configuration of electrodes on a subject's face for EOG signal acquisition.

Configuration of electrodes on a subject's face for EOG signal acquisition.

2.4. Validation Study

The validation study used the same picture sets and experimental procedure as the main study. In contrast to the main study, no eye movements were recorded from the participants. Instead, participants were asked for realtime feedback on whether or not each picture had been shown before by pressing two buttons on a keyboard. We collected feedback from eight participants other than those participating in the main study - four male and four female - aged between 21 and 28 years (M=24.3, SD=±3.8).

2.5. Eye Movement Analysis

Eye movements can be analyzed using EOG signals. Before signal processing a preprocessing is critical to suppress extra information from EOG signals. Signal processing techniques helps removing unnecessary information such as noise and baseline drift.

2.6. Noise and Baseline Drift Removal

Like other biological signals, EOG signals are often affected by noise and baseline drift. Noise in EOG signals may has several sources such as the residential power line, the measurement circuitry, electrodes, and wires, or other interfering physiological sources such as electromyographic (EMG) signals. To cope with these artifacts, first of all EOG signals were low-pass filtered with cutoff frequency at 30 Hz integrated in the device. Moreover an additional 50 Hz notch filter was applied to suppress the power line noise. The notch filter was applied using G.tec's Application Programming Interface (API) for MATLAB. This API contains commands which give full access to the amplifier. There are commands for reading the data, setting the band pass and notch filters, changing the sampling frequency of the amplifier, defining bipolar derivations and calibrating the system. Baseline drift is a slow signal change superposing the EOG signal but mostly unrelated to eye movements. It has many possible sources such as interfering background signals or electrode polarization. For baseline drift removal, firstly an approximated multilevel 1D wavelet decomposition at level twelve was performed using Daubechies wavelets on each EOG signal component. The reconstructed decomposition coefficients gave baseline drift estimation. Subtracting this estimation from each original signal component yielded the corrected signals with reduced drift offset.

2.7. Saccade Detection

In an earlier work, Bulling et al. introduced the Continuous Wavelet Transform - Saccade Detection (CWT-SD) algorithm (Bulling et al., 2011). Briefly, CWT-SD detects saccades by thresholding on the continuous 1-D wavelet coefficient vector computed from the de-noised and baseline drift removed HEOG and VEOG. CWT-SD takes physiological saccade characteristics into account to increase the robustness of detection (Duchowski, 2007).

2.8. Fixation Detection

Bulling et al. algorithm for fixation detection exploits the fact that fixation points tend to cluster together closely in time. Thus, by thresholding on the dispersion of these points, fixations can be detected (Widdel, 1984). Based on the output of the CWT-SD algorithm, dispersion and duration values are calculated for each nonsaccadic segment. If the dispersion is below a maximum threshold, and the duration above a minimum threshold, a fixation is detected.

2.9. Blink Detection

Blinks appear in vertical EOG signal component in the form of overshoots in signal amplitude. The Continuous Wavelet Transform - Blink Detection (CWT-BD) algorithm was used to detect blinks in VEOG. Similar to the algorithm for saccade detection, this algorithm uses thresholding of wavelet coefficients. In contrast to saccades, a blink is characterized by a short sequence of two large peaks in the coefficient vector, one positive and the other negative. The time between these peaks is much smaller than for saccades. Thus, blinks are distinguished from saccades by applying a maximum threshold on this time difference.

2.10. Feature Extraction and Selection

The two-class recognition problem of discriminating between pictures that were only seen once (class “nonrepeated”) and pictures that were seen twice (class “repeated”) by the participants were considered. After the removal of all eye movement data which belonged to the intervals, all picture instances (picture and corresponding eye movement data) of all single exposures to the “non-repeated” class, and picture instances of two times exposures were assigned to the “repeated” class. Feature extraction was run on all picture instances. 81 features were extracted of which most having statistical features of saccades, blinks and fixations (see Table 1). The features were calculated on both HEOG and VEOG.
Table 1

Feature descriptions

Feature No.Feature DescriptionFeature No.Feature Description
1Gaze Dispersion42Mean of saccade velocities in VEOG
2Mean of fixation duration43Median of saccade velocities in VEOG
3Gaze numbers44Variance of saccade velocities in VEOG
4Saccade numbers in HEOG45Maximum of saccade amplitudes in VEOG
5Mean of saccade amplitude in HEOG46Maximum of saccade durations in VEOG
6Median of saccade amplitudes in HEOG47Saccade numbers only in vertical direction
7Standard deviation of saccade amplitudes in HEOG48Entropy of HEOG signal
8Variance of saccade amplitudes in HEOG49Entropy of VEOG signal
9Mean of saccade durations in HEOG50Energy of HEOG signal
10Variance of saccade durations in HEOG51Energy of VEOG signal
11Mean of saccade velocities in HEOG52Energy of HEOG and VEOG multiplication
12Median of saccade velocities in HEOG53Autocorrelation of HEOG
13Variance of saccade velocities in HEOG54Autocorrelation of VEOG
14Maximum of saccade amplitudes in HEOG55Cross correlation between HEOG and VEOG
15Maximum of saccade durations in HEOG56Mutual information of HEOG with itself
16Saccade numbers only in horizontal direction57Mutual information of VEOG with itself
17Rightward saccade numbers58Mutual information of HEOG and VEOG
18Mean of rightward saccade amplitudes59Blink numbers
19Variance of rightward saccade amplitudes60Mean of blink amplitudes
20Mean of rightward saccade durations61Median of blink amplitudes
21Variance of rightward saccade durations62Mean of blink velocities
22Mean of rightward saccade velocities63Median of blink velocities
23Variance of rightward saccade velocities64Variance of blink velocities
24Leftward saccade numbers65Variance of blink amplitudes
25Mean of leftward saccade amplitudes66Mean of blink durations
26Variance of leftward saccade amplitudes67Variance of blink durations
27Mean of leftward saccade durations68Upward saccade numbers
28Variance of leftward saccade durations69Mean of upward saccade amplitudes
29Mean of leftward saccade velocities70Variance of upward saccade amplitudes
30Variance of leftward saccade velocities71Mean of upward saccade durations
31Saccade numbers in VEOG72Variance of upward saccade durations
32Mean of saccade amplitudes in VEOG73Mean of upward saccade velocities
33Median of saccade amplitudes in VEOG74Variance of upward saccade velocities
34Standard deviation of saccade amplitudes in VEOG75Downward saccade numbers
35Variance of saccade amplitudes in VEOG76Mean of downward saccade amplitudes
36Mean of saccade durations in VEOG77Variance of downward saccade amplitudes
37Variance of saccade durations in VEOG78Mean of downward saccade durations
38Up-Right saccade numbers79Variance of downward saccade durations
39Up-left saccade numbers80Mean of downward saccade velocities
40Down-Right saccade numbers81Variance of downward saccade velocities
41Down-Left saccade numbers
Feature descriptions For feature selection a filter scheme over the commonly used wrapper approaches was used because of the lower computational costs and thus shorter runtime. Minimum redundancy maximum relevance algorithm (Peng et al., 2005) for feature selection was used in this work. The mRMR algorithm selects a feature subset of arbitrary size S best characterizing the statistical properties of the given target classes based on the ground truth.

2.11. Classification and Performance Evaluation

A support vector machine (SVM) with a linear kernel was used for classification. All parameters of the saccade, fixation, and blink detection algorithms were fixed to values common to all participants. For evaluation, a leave-one-person-out scheme was followed, by which the datasets of all but one participant were combined and used for training (the “training set”). The dataset of the remaining participant was used for testing (the “test set”). This was repeated for each participant. Feature selection was performed solely on the training set.

3. Results

3.1. Results for Each Picture Category

On average, participants from the validation study were able to correctly identify pictures that had previously been shown with an accuracy of 76.3% for the first part of the experiment. This accuracy was 91.1% for soundmatched pictures during the second part of the experiment. Given the above high accuracy, in the following analysis it could be assumed that participants in the main study perfectly remembered most of the pictures which were already presented. Based on the data recorded in the main study, Figure 4 summarizes the overall recognition performance using person-independent parameters and training for each picture category and each part of the experiment including picture presentation with and without sounds. The bars contrast true positive rate (TPR) to false positive rate (FPR). Figure 4 demonstrates a considerable improvement in recognition performance for the second part of the experiment where pictures presented with their related sounds compared with the first part of the experiment wherein pictures were presented without sound. This improvement is re-demonstrated through ROC space in Figure 5. Recognition performance reduces if the points in ROC spaces of Figure 5 get closer to the up-left side.
Figure 4

Overall recognition performance for each picture category and each part of the experiment (including with and without sound) for all subjects.

Figure 5

Recognition performance improvement in participants from the first part of the experiment that pictures were presented without sound (blue crosses) to the second part wherein pictures were presented with their related sound (red circles).

Overall recognition performance for each picture category and each part of the experiment (including with and without sound) for all subjects. Recognition performance improvement in participants from the first part of the experiment that pictures were presented without sound (blue crosses) to the second part wherein pictures were presented with their related sound (red circles). Figures 6 and 7 show the range of differences in recognition performance for each individual participant respectively from the first to the second part of the experiment. A comparison between Figures 6 and 7 shows considerable reductions and decrease respectively in TPR and FPR for most of the participants in the four picture categories. These improvements are due to the auditory effect addition in the second part of the experiment.
Figure 6

Classification performance of repeated and non-repeated pictures without sound in the four categories for all of the subjects

Figure 7

Classification performance of repeated and non-repeated pictures with sound in the four categories for all of the subjects

Classification performance of repeated and non-repeated pictures without sound in the four categories for all of the subjects Classification performance of repeated and non-repeated pictures with sound in the four categories for all of the subjects

3.2. Further Analysis of the Faces Picture Category

Based on the cognition-aware memory assistant concept outlined in the introduction and the insights from the experiments done by Bulling and Roggen in 2011, the results for the faces picture category was analyzed in more detail.

3.3. Top discriminative EOG Features

Feature rankings using mRMR was analyzed on each of the twelve leave-one-person-out training sets for the faces category. The rank of a feature is the position at which mRMR selected it within a set. The position corresponds to the importance through which mRMR assesses a feature's ability to discriminate between classes in combination with the features already selected. Figures 8 and 9 show the top 16 features according to the median rank over all sets (see Table 1 for features’ description). For each feature, the vertical bar represents the spread of mRMR ranks for the twelve training sets. The most useful features are those found with the highest rank (close to one) for most training sets as indicated by shorter bars. As illustrated in Figures 8 and 9, the top discriminative features are mostly common for the two parts of the experiment.
Figure 8

Top 16 eye movement features selected by mRMR for all twelve training sets for the faces picture category. X-axis shows feature numbers and groups; the key on the right shows the corresponding feature names as described in Table 1; Y-axis shows the rank.

Figure 9

The top 16 eye movement features selected by mRMR for all twelve training sets for the faces picture category. X-axis shows feature numbers and groups; the key on the right shows the corresponding feature names as described in Table 1; Y-axis shows the rank.

Top 16 eye movement features selected by mRMR for all twelve training sets for the faces picture category. X-axis shows feature numbers and groups; the key on the right shows the corresponding feature names as described in Table 1; Y-axis shows the rank. The top 16 eye movement features selected by mRMR for all twelve training sets for the faces picture category. X-axis shows feature numbers and groups; the key on the right shows the corresponding feature names as described in Table 1; Y-axis shows the rank.

4. Conclusion

Our findings indicated that adding auditory information improves discrimination between familiar and unfamiliar pictures in participants. Moreover, this auditory effect leaves an impact on EOG signal patterns so that to make reduction in recognition performance.

5. Discussion

Recognizing human cognitive-context from biosignals would be so promising in pervasive computing. The findings of this study were in line with those by Bulling & Roggen, 2011 on memory assistant realization. However, utilizing a near-real world set up to achieve a robust and reliable memory assistant system is required to address remaining challenges. An important challenge for a real-world implementation is the co-influence of the task, situation, and cognitive processes on a subject's eye movements. In laboratory settings, these influences can be minimized by using a constrained experimental setup and well-defined visual stimuli. Meanwhile, the everyday settings can typically not be controlled in a similar fashion. It is therefore crucial to identify and separate these different sources of influence for robust recognition of visual memory recall and other cognitive processes. This problem could be addressed by using a multi-modal approach for context recognition and annotation that incorporates additional modalities to eye tracking, such as proximity sensors, GPS for localization, inertial measurement units for head movements, or eye contact sensors (Dickie et al., 2004). This leads to a second challenge. Personal encounters in daily life differ considerably from the situation investigated here. In these settings, facial expressions of conversational partners change continuously, the viewpoint is dynamic, and other visual stimuli may attract attention and lead to “random” saccades to other entities in the surrounding environment. In addition, personal encounters may range from longer face-to-face discussions between two people, over glances to faces of others while in transit, to looking at several faces of a group of people in succession. This will require advanced methods for robust detection of when and how people look at each other's face. One possible solution to this problem is to augment the analysis of eye movement dynamics – as presented here - with a computer vision system for face detection and a wearable gaze tracker to identify the points while a subjects lookes at a face (Bulling & Roggen, 2011). In the current experiment, participants were asked to to look at a large screen during which distinct eye movements were provoked and could easily be measured using EOG. It remains to be investigated whether current wearable eye trackers whether EOGor video-based - are accurate enough to capture eye movement characteristics which reflect visual memory recall processes on smaller screens (e.g. on a mobile phone) or with the person being in transit (Bulling & Roggen, 2011). It has been reported that someparticipants may feel bored or tired during such experiments (Bulling & Roggen, 2011). To address this problem we reduced the duration of picture presentation and the number of presented pictures in comparison to an earlier published report (Bulling & Roggen, 2011) thus, none of the participants reported that they get bored during the experiment. Another possible advantage of such modification was getting closer to real world situation. Such findings may also open up new approaches to design lie detection systems. EOG signals in combination with other biosignals would make reliable evidences for these systems. Due to the unobtrusive nature of wired biosignal acquisition systems, the application of cognition-aware system using biosignals may be seen quite feasible in this field.
  22 in total

1.  The effect of cognitive load on saccadic eye movements.

Authors:  E Stuyven; K Van der Goten; A Vandierendonck; K Claeys; L Crevits
Journal:  Acta Psychol (Amst)       Date:  2000-03

2.  Saccadic eye movements and cognition.

Authors: 
Journal:  Trends Cogn Sci       Date:  2000-01       Impact factor: 20.229

3.  Feature selection based on mutual information: criteria of max-dependency, max-relevance, and min-redundancy.

Authors:  Hanchuan Peng; Fuhui Long; Chris Ding
Journal:  IEEE Trans Pattern Anal Mach Intell       Date:  2005-08       Impact factor: 6.226

4.  Eye blink frequency during different computer tasks quantified by electrooculography.

Authors:  J H Skotte; J K Nøjgaard; L V Jørgensen; K B Christensen; G Sjøgaard
Journal:  Eur J Appl Physiol       Date:  2006-11-07       Impact factor: 3.078

5.  More efficient scanning for familiar faces.

Authors:  Jennifer J Heisz; David I Shore
Journal:  J Vis       Date:  2008-01-17       Impact factor: 2.240

Review 6.  Eye-movement models for arithmetic and reading performance.

Authors:  P Suppes
Journal:  Rev Oculomot Res       Date:  1990

7.  Where we look when we steer.

Authors:  M F Land; D N Lee
Journal:  Nature       Date:  1994-06-30       Impact factor: 49.962

8.  Blinks and saccades as indicators of fatigue in sleepiness warnings: looking tired?

Authors:  R Schleicher; N Galley; S Briest; L Galley
Journal:  Ergonomics       Date:  2008-07       Impact factor: 2.778

9.  Visual fixation patterns during viewing of naturalistic social situations as predictors of social competence in individuals with autism.

Authors:  Ami Klin; Warren Jones; Robert Schultz; Fred Volkmar; Donald Cohen
Journal:  Arch Gen Psychiatry       Date:  2002-09

10.  Human gaze control during real-world scene perception.

Authors:  John M Henderson
Journal:  Trends Cogn Sci       Date:  2003-11       Impact factor: 20.229

View more
  4 in total

1.  Eye movement characteristics reflected fatigue development in both young and elderly individuals.

Authors:  Ramtin Zargari Marandi; Pascal Madeleine; Øyvind Omland; Nicolas Vuillerme; Afshin Samani
Journal:  Sci Rep       Date:  2018-09-03       Impact factor: 4.379

2.  An oculometrics-based biofeedback system to impede fatigue development during computer work: A proof-of-concept study.

Authors:  Ramtin Zargari Marandi; Pascal Madeleine; Øyvind Omland; Nicolas Vuillerme; Afshin Samani
Journal:  PLoS One       Date:  2019-05-31       Impact factor: 3.240

3.  Developmental Eye Movement (DEM) Test Norms for Mandarin Chinese-Speaking Chinese Children.

Authors:  Yachun Xie; Chunmei Shi; Meiling Tong; Min Zhang; Tingting Li; Yaqin Xu; Xirong Guo; Qin Hong; Xia Chi
Journal:  PLoS One       Date:  2016-02-16       Impact factor: 3.240

4.  Eye-Tracker Study of Influence of Affective Disruptive Content on User's Visual Attention and Emotional State.

Authors:  Anna Lewandowska; Izabela Rejer; Kamil Bortko; Jarosław Jankowski
Journal:  Sensors (Basel)       Date:  2022-01-11       Impact factor: 3.576

  4 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.