Literature DB >> 35769696

A Deep Learning Model for Three-Dimensional Nystagmus Detection and Its Preliminary Application.

Wen Lu1, Zhuangzhuang Li1, Yini Li1, Jie Li1, Zhengnong Chen1, Yanmei Feng1, Hui Wang1, Qiong Luo1, Yiqing Wang2, Jun Pan2, Lingyun Gu2, Dongzhen Yu1, Yudong Zhang3, Haibo Shi1, Shankai Yin1.   

Abstract

Symptoms of vertigo are frequently reported and are usually accompanied by eye-movements called nystagmus. In this article, we designed a three-dimensional nystagmus recognition model and a benign paroxysmal positional vertigo automatic diagnosis system based on deep neural network architectures (Chinese Clinical Trials Registry ChiCTR-IOR-17010506). An object detection model was constructed to track the movement of the pupil centre. Convolutional neural network-based models were trained to detect nystagmus patterns in three dimensions. Our nystagmus detection models obtained high areas under the curve; 0.982 in horizontal tests, 0.893 in vertical tests, and 0.957 in torsional tests. Moreover, our automatic benign paroxysmal positional vertigo diagnosis system achieved a sensitivity of 0.8848, specificity of 0.8841, accuracy of 0.8845, and an F1 score of 0.8914. Compared with previous studies, our system provides a clinical reference, facilitates nystagmus detection and diagnosis, and it can be applied in real-world medical practices.
Copyright © 2022 Lu, Li, Li, Li, Chen, Feng, Wang, Luo, Wang, Pan, Gu, Yu, Zhang, Shi and Yin.

Entities:  

Keywords:  benign paroxysmal positional vertigo; deep learning; neural network; nystagmus detection; vertigo

Year:  2022        PMID: 35769696      PMCID: PMC9236194          DOI: 10.3389/fnins.2022.930028

Source DB:  PubMed          Journal:  Front Neurosci        ISSN: 1662-453X            Impact factor:   5.152


Introduction

Of all the symptoms encountered clinically, vertigo is one of the most common complaints. Vertigo has a considerable impact on personal quality of life, which is exacerbated by aging (Neuhauser et al., 2005; Murdin and Schilder, 2015; Tonsen et al., 2016; Alyono, 2018). With a 12-month prevalence of 15–20% (Neuhauser, 2016), vertigo imposes a huge economic burden on primary health care with costs totalling 61.3 million pounds annually (Tyrrell et al., 2016; Kovacs et al., 2019). Unfortunately, the variety and heterogeneity of vestibular disorders greatly increases the difficulty in making a clinical diagnosis, leading to numerous repeat medical consultations with low rates of specific diagnoses (20–60%) and poor specialist referral rates (3–4%) (Kruschinski et al., 2008; Maarsingh et al., 2010a,b; Neuhauser, 2016). Unlike other diseases, vestibular disorders are difficult to diagnose due to the lack of typical signs and features. Nystagmus, an involuntary, rapid, rhythmic, oscillatory eye movement, is the most important sign for the differential diagnosis of vestibular disorders (Eggers et al., 2019). There are three directions of nystagmus: horizontal, vertical, and torsional. Its detection is widely used in routine clinical evaluation of patients with vertigo in specialty clinics, via visual observation with the naked eye or video nystagmography (VNG) (Bhansali and Honrubia, 1999; Eggert, 2007). However, nystagmus recognition poses many challenges in modern clinical practice, including a lack of specialists and medical resources, complex and heterogeneous characteristics that are difficult to analyse, and sensitivity limitations in nystagmus recognition by the naked eye, especially when the nystagmus is subtle. In practice, it is difficult to evaluate patients with droopy eyelids or eyelashes covering their pupils using VNG, and the interference of infrared light and cosmetics around the eyes can make it worse (Ganança et al., 2010). Therefore, establishing a model for three-dimensional (3D) nystagmus detection is of great urgency. With the advances in science and technology, a system for nystagmus detection could be achieved using artificial intelligence (AI). AI is an interdisciplinary subject dedicated to data-driven empirical learning (Wainberg et al., 2018) which has been considered a potential solution to several medical diagnostic challenges, especially in the fields of radiology and pathology. The accessibility, growth potential, and limited cost make AI a promising option for dealing with the lack of medical resources and specialists. The convolutional neural network (CNN) is one of the most widely used deep learning algorithms in AI-based applications, contributing to object classification, detection, and segmentation. This makes CNN promising for AI-based recognition of nystagmus due to its ability to capture specific features, extensive open-source codes, and the advanced research foundation for eye-tracking from other medical fields. However, pioneering research on CNN-based automated nystagmus detection encountered difficulty with pupil detection in some situations and fails to capture twisting eyeball movement. Still, there are proven object detection models (Szegedy et al., 2015, 2017) that have shown good performance in coping with the high frequency of noise (e.g., eye blink, head movements) faced in clinical practice. Currently, no nystagmus recognition system has been used for clinical diagnosis. Benign paroxysmal positional vertigo (BPPV) is a common cause of vertigo and is diagnosed in 17–42% of patients with vertigo (Schappert, 1992; Katsarkas, 1999; Hanley et al., 2001). The BPPV diagnostic procedure costs approximately 2,000 USD, and 65% of patients undergo unnecessary diagnostic tests or therapeutic interventions (Wang et al., 2014). Since BPPV can be easily cured once correctly diagnosed, such a waste of resources could be avoided with an advanced diagnostic strategy. Notably, BPPV allows for different types of nystagmus to be observed in specific head positions and its characteristic nystagmus is relatively easy to analyse, offering a low threshold for AI diagnosis. In this study, we developed an automatic system for detecting 3D eye movements based on deep learning. To improve the robustness and establish a reliable AI diagnostic system, we developed a new method to locate pupils accurately and detect iris twist. The model was validated in patients with BPPV and achieved high sensitivity and accuracy in nystagmus detection and disease diagnosis. In this study, we applied our model in BPPV diagnosis, not only as a real-world performance test of our algorithm model, but also in an attempt to develop an intelligent diagnostic system with real-world application potential.

Materials and Methods

We enrolled patients from the outpatient clinic of the Department of Otolaryngology-Head and Neck Surgery of the Sixth People’s Hospital of Shanghai Jiao Tong University between September 2017 and November 2021 who underwent vestibular function tests using infrared video goggles (Verti Goggles-M, ZEHNIT Medical Technology, Shanghai, China). All patients complaining of symptoms of vertigo or dizziness would undergo two positional tests: the supine Roll Test and Dix-Hallpike manoeuvre. Each test lasted for at least 30 s until the end of eye movement. The recorded videos were labelled by three experts based on the BPPV diagnostic criteria of the Bárány Society (Von Brevern et al., 2015; Yao et al., 2018). The datasets for validation were selected with similar proportions of positive samples from the training data to avoid biased evaluation. The input data of our model were comprised of 1–4 clinical videos of each patient.

Ethics Committee Approval

The study was approved by the Ethics Committee of Shanghai Sixth People’s Hospital and was conducted according to the Declaration of Helsinki. Written informed consent was obtained from all the participants. The study was registered in the Chinese Clinical Trials Registry (ChiCTR.org.cn) under the identifier ChiCTR-IOR-1711-506.

Experimental Procedure

The overall framework of our diagnostic system is shown in Figure 1. Portable video goggles were adopted to capture pupil movement during the Dix-Hallpike manoeuvre and supine roll test. The procedure consisted of four parts: pupil detection, iris torsion measurement, the deep learning model, and disease inference.
FIGURE 1

Framework of the automatic nystagmus detection system. Procedures of our auto diagnosis system: pupil locator system, iris torsion measure, data pre-processing, CNN-based nystagmus detection model, and disease inference. CNN, convolutional neural network.

Framework of the automatic nystagmus detection system. Procedures of our auto diagnosis system: pupil locator system, iris torsion measure, data pre-processing, CNN-based nystagmus detection model, and disease inference. CNN, convolutional neural network.

Pupil Locator

The raw clinic videos do not mark the position of the pupil centre; thus, the first step is to locate the pupil inside each frame. A pupil location algorithm was applied to locate the pupil centre in each video. Previous studies (Santini et al., 2018; Eivazi et al., 2019) have attempted to predict the parameters of pupil location using deep learning models. As the performance of deep learning algorithms continues to improve, pupil detection algorithms are often driven by data. Such data-driven models require a large amount of qualified data labelled by specialists, which is an expensive, slow, and error-prone manual process. To reduce the error and cost caused by annotation, we trained one deep-learning model with the architecture of Inception V4 (Szegedy et al., 2015, 2017) on an open-source dataset containing 66 high-quality, high-speed videos (Tonsen et al., 2016), and then used the pre-trained model to label our raw videos (Figure 2).
FIGURE 2

Objection detection model for pupil location. (A) Model architecture. (B) Feature extraction with different convolution kernels. (C) Visualisation of the pupil parameters: dot – pupil centre; circle – outer radius.

Objection detection model for pupil location. (A) Model architecture. (B) Feature extraction with different convolution kernels. (C) Visualisation of the pupil parameters: dot – pupil centre; circle – outer radius.

Torsional Movement Detection

Torsional nystagmus is a deterministic signal in BPPV diagnosis. Several methods have been proposed to measure the torsional movement (Wolberg and Zokai, 2000; Ojansivu and Heikkila, 2007; Alba et al., 2013) such as tracking stable iris features, template matching, and optical flow. We decided to use phase correlation techniques as our torsional measurement method, an approach that has typically been applied in image registration (Araujo and Dias, 1997; Abdullah-Al-Wadud et al., 2007). The circular iris pattern was transformed by log-polar transformation. An image augmentation technique called histogram equalisation (Abdullah-Al-Wadud et al., 2007) was applied to extract slight features on iris pattern. Phase correlation, a frequency domain technique with broad applications in image alignment (Reddy and Chatterji, 1996), was used for the estimation of the similarity measure of two images (Figure 3).
FIGURE 3

Iris torsion measure. (A) Iris extraction. Circles: iris boundaries; rectangles: log-polar (left) and linear-polar transform (right). (B) Original and equalised histogram. (C) Iris patterns before and after equalisation. (D) Phase-only correlation function.

Iris torsion measure. (A) Iris extraction. Circles: iris boundaries; rectangles: log-polar (left) and linear-polar transform (right). (B) Original and equalised histogram. (C) Iris patterns before and after equalisation. (D) Phase-only correlation function.

Log-Polar Transformation

The log-polar coordinate parameters ρ and θ denote the logarithmic radial distance from the pupil centre and the angle, respectively, corresponding to the radial distance from the centre and angle from the centre, respectively. Any point (x, y) in the original Cartesian plane can be reflected in the rectangular iris pattern (see computation below): (X0, Y0) and (Xnew, Ynew) correspond to the centre of the pupil and coordinate mapping from the Cartesian domain to the rectangular iris pattern, respectively. Formulas (1) and (2) are used to calculate the log-polar coordinates of each point (x, y) in the Cartesian plane, and formulas (3)-(6) are used to resize the log-polar coordinates: Where cols and rows represent the scale of the image, and Rmax denotes the maximum radius sampled from the image.

Histogram Equalisation

We first calculated the probability mass function of all the pixels in the grayscale histogram of the original image using formula (7), where k represents the grayscale value and N represents the total number of pixels with grayscale k. Second, formula (8) generates the discrete cumulative distributive function. The transformed grayscale histogram is generated by formula (9). The contrast of the iris pattern was enhanced using histogram equalisation, which is useful for the further application of the phase correlation method.

Phase Correlation

Several properties of Fourier transform, such as translation, rotation, reflection, and scale in the frequency domain, have been exploited for image registration. Phase correlation relies on the translation property of Fourier transform, based on estimating the shift between two images by calculating the maximum of the phase-only correlation function, which is defined as the inverse FFT of the normalized cross-spectrum between two images. Let f and g be the pixel signals of two images with displacement < dx, dy >, that is: Let the Fourier Transform function be: The corresponding relationship of F and G is given by: According to the properties of Fourier transform, the translational movement of the time-domain signal can be expressed by the phase difference in the frequency domain, which is equivalent to the phase of the cross-power spectrum: The inverse Fourier transform of the phase difference is an impulse function, and the peak location calculated from formula (15) is the point of image registration. The sharp peak appears only when two images have best matched the height which gives a similarity measure for image alignment. Also, the location of the peak represents the displacement between images, which is what we need to quantify the eye torsional movement between two consecutive frames. In horizontal/vertical movements, the velocity curve was obtained by applying differencing to the time series of pupil centre coordinates, while the velocity in torsional dimensions was defined as the similarity measure between two consecutive frames (Figure 4). Due to the existence of noise frames in test videos caused by eye blinking, head shifting, and mistaken operations, we applied the DBSCAN method (an unsupervised machine learning algorithm that has excellent performance for detecting outliers) to eliminate the noise data.
FIGURE 4

Eye movement data generation. Each video was transformed to eye movement velocity in three dimensions. Horizontal and vertical velocity: coordinates of the pupil centre. Torsional velocity: the shift between two consecutive frames.

Eye movement data generation. Each video was transformed to eye movement velocity in three dimensions. Horizontal and vertical velocity: coordinates of the pupil centre. Torsional velocity: the shift between two consecutive frames.

Deep Learning Model

We adopted several data augmentation technologies to enhance our model’s performance due to the insufficient amount of data for deep learning model training. We first split clinic videos into overlapping sub-samples with a fixed length (Figure 5A). Then, each video clip was horizontally and vertically reversed, and white noise data was added (Figure 5B). Finally, the over-sampling technique was applied to balance the labelled data in nystagmus and non-nystagmus. One-dimensional (1D CNN) was used as the architecture of our nystagmus detection model, an architecture that performed well in similar tasks (Yildirim et al., 2018; Zabihi et al., 2019). Patients labelled positive in the supine roll test showed horizontal nystagmus, while the Dix-Hallpike manoeuvre was applied to detect vertical/torsional nystagmus. We trained three models for horizontal, vertical, and torsional nystagmus detection (Figure 5C).
FIGURE 5

Data pre-processing and the deep learning model for nystagmus detection. (A) The rolling cut of the velocity curve to fix the length of the time series; sub-samples including a nystagmus pattern (marked as red) are labelled as positive (otherwise negative). (B) Data augmentation methods applied to generate new examples of nystagmus. Upper left: Original data with nystagmus signals. Upper right: Data flipped on the x-axis. Bottom left: Data flipped on the y-axis. Bottom right: Add white noise (C) shows the model structures. GMP, global max pooling; MLP, multi-layer perception.

Data pre-processing and the deep learning model for nystagmus detection. (A) The rolling cut of the velocity curve to fix the length of the time series; sub-samples including a nystagmus pattern (marked as red) are labelled as positive (otherwise negative). (B) Data augmentation methods applied to generate new examples of nystagmus. Upper left: Original data with nystagmus signals. Upper right: Data flipped on the x-axis. Bottom left: Data flipped on the y-axis. Bottom right: Add white noise (C) shows the model structures. GMP, global max pooling; MLP, multi-layer perception.

Disease Inference

The variety and complexity of BPPV diseases increased the difficulty of automatic diagnosis. A decision tree procedure was thus established to determine the exact type of BPPV disease (Figure 6C), depending on the direction and duration of the nystagmus signal.
FIGURE 6

Disease diagnosis process. (A) Peak detection: all velocity peaks in two directions are detected. (B) The predicted labels of test data, longest consecutive positive sub-samples represent the position of nystagmus. (C) The decision tree that simulates the diagnosis of specialists.

Disease diagnosis process. (A) Peak detection: all velocity peaks in two directions are detected. (B) The predicted labels of test data, longest consecutive positive sub-samples represent the position of nystagmus. (C) The decision tree that simulates the diagnosis of specialists.

Nystagmus Direction

The velocity peaks are detected from the time series (Figure 6A), then the median velocity is obtained in two directions. The direction with a larger absolute median value is determined to be the direction of nystagmus.

Nystagmus Duration

The prediction steps label each sub-sample in the test set, and the length of the longest consecutive positive segmentation is defined as the duration of nystagmus. Figure 6B shows an example. The test video is separated into 31 sub-samples for model prediction and the number of consecutive positive samples given by model prediction is 19; an estimate of nystagmus duration can be calculated with formula (16): Where frame is the length of sub-samples (set as 400/600); N is the number of consecutive positive sub-samples; and overlap is the length of the overlapped part.

Statistical Analyses

The sample size was determined by the total amount of high-quality labelled data collected. The lack of available data was the most significant challenge when applying deep learning algorithms for BPPV diagnosis; therefore, we used all the data to strengthen the performance, instead of performing a power analysis. The 854 patient cases were randomly selected from thousands of patients in the case pool and split into training, validation, and testing datasets by simple random sampling. The testing dataset was not available during modelling to ensure that the experimenters were blind to outcome assessment. Statistical analysis was performed separately for the nystagmus detection model and BPPV disease inference to evaluate the overall performance of our automatic diagnosis system. For the primary datasets (training, validation, and test), binary classification evaluation was conducted by calculating the accuracy and area under the receiver operating characteristic curve (AUC) (which provides an aggregate measure of model performance at all classification thresholds) for each model in three directions. For disease inference, we computed the true positive (TP), false positive (FP), false negative (FN), true negative (TN), precision, recall accuracy, and F1 score at binary decision thresholds for the aggregate measure of the inference performance on different types of BPPV diseases.

Results

Participant Characteristics

We enrolled a total of 854 patients from the outpatient clinic of the Department of Otolaryngology-Head and Neck Surgery of the Sixth People’s Hospital of Shanghai Jiao Tong University between September 2017 and November 2021 who underwent vestibular function tests using infrared video goggles. We collected clinical videos from these patients’ records, including 3,496 horizontal movements and 5,962 vertical/torsional movements (Table 1). Among the 854 patients in our dataset, 304 (35.6%) were randomly selected as the training set, 93 (10.9%) were randomly selected as the validation set, 122 (14.3%) were selected as the testing set for nystagmus model performance evaluation, and the remaining 457 (53.5%, including the testing set previously mentioned) were selected to evaluate the accuracy of disease inference. To avoid data leaks, the split of training and validation datasets were determined by the patients, not clinical videos.
TABLE 1

Summary of the data sets (baseline).

Sex (M/F)Age (Mean ± SD)
Training
LP12/2054.47 ± 5.28
RP11/3858.03 ± 13.44
LH4/855.50 ± 17.28
RH4/1655.65 ± 15.21
LH cu3/261.20 ± 12.79
RH cu1/253.67 ± 27.43
Negative54/12947.52 ± 16.69
Total89/21551.16 ± 16.61
Validation
LP0/1059.30 ± 15.71
RP2/653.75 ± 16.18
LH1/344.75 ± 12.61
RH2/847.40 ± 14.32
LH cu0/138.00
RH cu0/278.50 ± 14.85
Negative16/4250.84 ± 15.42
Total21/7251.83 ± 15.74
Testing
LP19/2452.00 ± 15.55
RP31/6953.60 ± 14.65
LH4/2154.12 ± 18.32
RH19/3657.23 ± 15.58
LH cu6/1052.06 ± 14.38
RH cu7/346.20 ± 19.70
Negative61/14748.53 ± 18.09
Total147/31051.39 ± 16.98

L, left; R, right; P, posterior semi-circular canal; H, horizontal semi-circular canal; cu, cupulolithiasis; M, male; F, female; SD, standard deviation.

Summary of the data sets (baseline). L, left; R, right; P, posterior semi-circular canal; H, horizontal semi-circular canal; cu, cupulolithiasis; M, male; F, female; SD, standard deviation.

Model Performance

Our model was trained to predict nystagmus in each single-frame segment. The model performance for identifying different types of nystagmus is summarised in Table 2. The AUC and accuracy of the horizontal and torsional models are shown in Figure 7. The overall performance of 502 cases from 457 patients (some cases taken by the same patient at a different visiting time) in terms of disease inference is shown in Tables 3, 4. The torsional model achieved a sensitivity and specificity of 0.8848 and 0.8841, respectively.
TABLE 2

Model performance in detecting horizontal, torsional, and vertical nystagmus.

HorizontalTorsionalVertical
Cases11412516
Samples15,92020,8161,882
AUC0.98250.95740.893
ACC0.93030.87950.905

AUC, area under curve; ACC, accuracy.

FIGURE 7

Model performance. The receiver operating characteristic curve (ROC) of model performance classifying nystagmus types after model training. (A) The area under the ROC for measuring horizontal nystagmus is 0.982. (B) The area under the ROC for measuring torsional nystagmus is 0.957.

TABLE 3

One-vs-rest multi-class prediction results after symptoms inference.

Model prediction

NegativeLPRPLH-caLH-cuRH-caRH-cu
Doctor’s diagnosisNegative206980316
LP73600000
RP144890010
LH-ca60121051
LH-cu01101402
RH-ca40040451
RH-cu00001110

L, left; R, right; P, posterior semi-circular canal; H, horizontal semi-circular canal; ca, canalolithiasis; cu, cupulolithiasis.

TABLE 4

Summary results of the model in diagnosing types of benign paroxysmal positional vertigo (BPPV).

NumberTPR/ RecallFPRACCTNRPrecisionF1-scores
5020.88480.11590.88450.88410.89810.8914

ACC, accuracy; TPR (sensitivity), true positive rate; FPR, false-positive rate; TNR (specificity), true negative rate.

Model performance in detecting horizontal, torsional, and vertical nystagmus. AUC, area under curve; ACC, accuracy. Model performance. The receiver operating characteristic curve (ROC) of model performance classifying nystagmus types after model training. (A) The area under the ROC for measuring horizontal nystagmus is 0.982. (B) The area under the ROC for measuring torsional nystagmus is 0.957. One-vs-rest multi-class prediction results after symptoms inference. L, left; R, right; P, posterior semi-circular canal; H, horizontal semi-circular canal; ca, canalolithiasis; cu, cupulolithiasis. Summary results of the model in diagnosing types of benign paroxysmal positional vertigo (BPPV). ACC, accuracy; TPR (sensitivity), true positive rate; FPR, false-positive rate; TNR (specificity), true negative rate.

Discussion

In this study, we created a multidimensional BPPV diagnosis system with a sensitivity of 0.8848 and specificity of 0.8841, based on deep learning models. Poor specialist referral rates and the limited sensitivity of nystagmus detection with the naked eye have led to the delayed diagnosis and mismanagement of vertigo, which may both significantly impact individual health and a heavy burden on primary care (Lopez-Escamez et al., 2005; Wang et al., 2014).

Previous Work

Previous studies have investigated the automatic detection of nystagmus, while an entire AI-based BPPV diagnosis system has not been implemented. Zhang et al. (2021) proposed a model for torsional BPPV nystagmus based on optical flow techniques which could effectively avoid the disturbance due to eyelash occlusion and pupil deformation. However, this model only supplied a basal framework for torsional nystagmus detection and could not be directly applied in disease diagnosis. Lim et al. (2019) developed a diagnostic decision support system for BPPV diagnosis using a 2D-CNN model. They showed that this system could predict the affected canals with high sensitivity and specificity with a large amount of training data, while this was limited when annotated data by otologic experts was insufficient. Newman et al. (2021) proposed an 1D-CNN model to predict nystagmus from corner-retinal potentials captured by the continuous ambulatory vestibular assessment (CAVA) device. This method was annotative and effective; however, it is not feasible for torsional nystagmus and also not acceptable for short-time positional tests (since patients have to wear CAVA device for a long time).

Improvements

Our study was specifically designed to address the limitations of these previous studies. Thus, we generated a complete system that refined eye movement velocity curves from raw clinic test videos, trained deep learning models to predict horizontal/vertical/torsional nystagmus, and automatically diagnosed BPPV diseases using quantitative metrics. Moreover, our system is interpretable, and the accessibility of data (e.g., eye movement time-series, torsional movement images, quantitative metrics) generated in each procedure of our system is important in the medical field. The BPPV detection system developed in our study can automatically detect horizontal/torsional/vertical nystagmus and BPPV diseases with a high AUC, F1-score, sensitivity, and specificity. The AUC is an overall measure of accuracy that combines sensitivity and specificity into a single metric. Our nystagmus detection model obtained good AUCs in both horizontal and torsional directions (Table 2). However, in terms of automatic BPPV diagnosis, the AUC was not reliable because of the class imbalances in our patient distribution. Data are said to be class imbalanced when the class distributions are highly imbalanced. For these multi-class cases, without loss of generality, the minority class is usually very infrequent. If one applies traditional classifiers on the dataset, the model would prefer to predict everything as negative (majority class), which was regarded as a serious problem in learning from highly imbalanced datasets. Therefore, F1 scores, sensitivity, and specificity are more suitable for sparse multi-label situations. We found a sensitivity of 0.8848, specificity of 0.8841, and F1 score of 0.8914 in final BPPV diagnosis (Tables 3, 4). The improvement of our torsional movement detection method, compared with previous work, (Ong and Haslwanter, 2010; Jin et al., 2020; Zhang et al., 2021) can be attributed to the implementation of several image processing techniques. We first adopted the log-polar transformation to extract iris features, then applied phase correlation techniques to measure the shift between each frame. Subsequently, the torsional nystagmus signals could be obtained from the movement pattern of the iris, which is acceptable for deep learning model training. Techniques mentioned in previous studies (Ong and Haslwanter, 2010; Jin et al., 2020; Zhang et al., 2021) belong to two main categories. The first type of technique involves extraction of iris feature points and then using them to calculate the displacement of the iris based on the correspondence of the feature points between the two frames of images, such as the optical flow method. The second type of technique involves the use of a feature similarity comparison, such as the Oriented FAST and Rotated BRIEF algorithm, image pixel histogram comparison, and template matching, to find the displacement angle of the iris when the maximum similarity occurs based on the comparison information obtained between the target and the reference image. However, feature extraction is essentially the extraction of texture information of the iris, which is highly susceptible to the influence of eyelids, reflections, and other factors. Therefore, manual processing of noise points is required, and this manual scheme cannot be applied to frame-by-frame recognition of videos. Our method makes two improvements to the previous algorithms based on template matching. First, the logarithmic polar coordinate transform is used instead of the original linear polar coordinate transform, which can better restore the iris texture features. Second, the phase correlation method is used instead of the original stencil-matching method. The phase correlation method can provide an offset in two dimensions and focus more on extracting the overall iris information. It is more robust than pixel-level schemes such as template matching and optical flow. Another technical improvement in our study was the implementation of a deep learning model in pupil detection, based on previous work. General approaches for pupil detection rely on computer vision methods such as edge detection, intensity thresholding, and intensity gradient distribution, which are not feasible here due to the irregular eye movement, fast head-turning, and image processing challenges (e.g., reflections, eyelid closure, video blurring). Some previous work (Santini et al., 2018; Eivazi et al., 2019) shows the outstanding performance on pupil detection using deep learning models. In this study, we applied an object detection model to localise a patient’s pupil in a clinical video. We pre-trained a pupil detection model on a public pupil dataset (Tonsen et al., 2016) and then transferred the model to our private datasets. The diagnosis of BPPV is based on the different characteristics of the nystagmus elicited by the provoking manoeuvres (Pérez-Vázquez and Franco-Gutiérrez, 2017). The comprehensive diagnosis of BPPV includes the specification of the affected semi-circular canal(s) and pathophysiology (canalolithiasis or cupulolithiasis) (Von Brevern et al., 2015). Our model can identify different types of nystagmus and provide the complete diagnosis of BPPV, although some rare variants of BPPV, such as canalolithiasis of the anterior canal (AC-BPPV) and cupulolithiasis of the posterior canal (PC-BPPV-cu), were not presented in our results due to the limited number of patients with these diseases. Since the model achieved high accuracy and sensitivity in BPPV diagnosis, the automated diagnostic support system based on our model could greatly benefit BPPV patients in primary care practice or emergency departments. Moreover, our model can be widely applied in various scenarios when embedded in mobile phones or other devices for eye tracking. As for the target population and clinical application, we attempted to identify nystagmus in all patients with vertigo, not only those with BPPV disease (Lim et al., 2019; Newman et al., 2021; Zhang et al., 2021). Nystagmus and other nystagmus-like movements are important signals for identifying whether a patient has vestibular and neurological disorders (Eggers et al., 2019). Detailed examinations of eye movements have been shown to play a key role in differentiating between central and peripheral vertigo (Kattah et al., 2009; Brandt and Dieterich, 2017; Pudszuhn et al., 2020). Our system can identify irregular non-BPPV nystagmus, including upbeat and downbeat nystagmus, indicating different pathologies and helping in clinical diagnosis and treatment. Additionally, we can extract and quantify all parameters of nystagmus, which allows for the objective assessment of prognosis and therapeutic efficacy in patients with vertigo. Considering its complexity and pronounced individual heterogeneity, our system could provide new insights into subtype identification. It is worth noting that this system can only provide a reference diagnosis for limited categories of vertigo, raising the need for further understanding of vertigo pathophysiology and a comprehensive system with various clinical characteristics. Thus, our system provides a solid basis for further research.

Limitations

Our study has several limitations. First, this system was developed and optimised with data obtained from a single centre; thus, it requires further validation and optimisation based on large-scale data from multiple centres in the future. Second, there are currently no standardised algorithms or open-source datasets for BPPV diagnosis, meaning that it is difficult to set a benchmark for model evaluation. Moreover, due to the privacy restrictions of clinical data, it is problematic for our model to be widely tested on other datasets. The third limitation was the DC drift of the signal. Positional tests usually last for several minutes, with the continuous movement of the patient’s pupil. We applied several methods to counteract this, including using the velocity signal instead of displacement, reducing the input length of the time-series, and consistently tracing the pupil centre as the coordinate origin. However, the accuracy of the model is still affected, especially in torsional nystagmus detection.

Conclusion

In summary, we developed an automated, interpretable, and validated system that performs real-time video quality feedback, pupil location, iris torsion measurement, data augmentation, nystagmus detection, and disease inference. With these functions, the system can provide a clinical reference and facilitate nystagmus detection and diagnosis. Hence, the proposed method can be applied in real-world medical practice.

Data Availability Statement

The raw data supporting the conclusions of this article will be made available by the authors, without undue reservation.

Ethics Statement

The studies involving human participants were reviewed and approved by the Ethics Committee of Shanghai Sixth People’s Hospital. The patients/participants provided their written informed consent to participate in this study.

Author Contributions

DY, YZ, and HS contributed to the conceptualisation, funding acquisition, and writing—review and editing. ZC, YF, HW, QL, and JL contributed to the investigation and resources. YW, JP, and LG contributed to the software. WL, ZL, and YL contributed to the writing – original draft, methodology, project administration, and validation. SY supervised the study. All authors contributed to the article and approved the submitted version.

Conflict of Interest

YW, JP, and LG were employed by IceKredit Inc. The remaining authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Publisher’s Note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.
  30 in total

Review 1.  A systematic review of vertigo in primary care.

Authors:  K Hanley; T O'Dowd; N Considine
Journal:  Br J Gen Pract       Date:  2001-08       Impact factor: 5.386

2.  National Ambulatory Medical Care Survey: 1989 summary.

Authors:  S M Schappert
Journal:  Vital Health Stat 13       Date:  1992-04

Review 3.  Eye movement recordings: methods.

Authors:  Thomas Eggert
Journal:  Dev Ophthalmol       Date:  2007

Review 4.  The epidemiology of dizziness and vertigo.

Authors:  H K Neuhauser
Journal:  Handb Clin Neurol       Date:  2016

5.  Benign paroxysmal positional vertigo (BPPV): idiopathic versus post-traumatic.

Authors:  A Katsarkas
Journal:  Acta Otolaryngol       Date:  1999       Impact factor: 1.494

6.  [Acute vestibular syndrome in emergency departments : Clinical differentiation of peripheral and central vestibulopathy].

Authors:  A Pudszuhn; A Heinzelmann; U Schönfeld; S M Niehues; V M Hofmann
Journal:  HNO       Date:  2020-05       Impact factor: 1.284

7.  Epidemiology of vestibular vertigo: a neurotologic survey of the general population.

Authors:  H K Neuhauser; M von Brevern; A Radtke; F Lezius; M Feldmann; T Ziese; T Lempert
Journal:  Neurology       Date:  2005-09-27       Impact factor: 9.910

8.  Use of the Bárány Society criteria to diagnose benign paroxysmal positional vertigo.

Authors:  Qingxiu Yao; Hui Wang; Qiang Song; Haibo Shi; Dongzhen Yu
Journal:  J Vestib Res       Date:  2018       Impact factor: 2.435

9.  HINTS to diagnose stroke in the acute vestibular syndrome: three-step bedside oculomotor examination more sensitive than early MRI diffusion-weighted imaging.

Authors:  Jorge C Kattah; Arun V Talkad; David Z Wang; Yu-Hsiang Hsieh; David E Newman-Toker
Journal:  Stroke       Date:  2009-09-17       Impact factor: 7.914

10.  Dizziness reported by elderly patients in family practice: prevalence, incidence, and clinical characteristics.

Authors:  Otto R Maarsingh; Jacquelien Dros; François G Schellevis; Henk C van Weert; Patrick J Bindels; Henriette E van der Horst
Journal:  BMC Fam Pract       Date:  2010-01-11       Impact factor: 2.497

View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.