STUDY OBJECTIVES: To assess the accuracy of an automated system (Morpheus I Sleep Scoring System) for analyzing and quantifying polysomnographic data from a population with sleep-disordered breathing. SETTING: Sleep laboratory affiliated with a tertiary care academic medical center. MEASUREMENTS AND RESULTS: 31 diagnostic polysomnograms were blindly analyzed prospectively with the investigational automated system and manually by 2 registered polysomnography technologists (M1 & M2) from the same laboratory. Sleep stages, arousals, periodic limb movements, and respiratory events (apneas and hypopneas) were scored by all 3. Agreement, Cohen kappa, and intraclass correlation coefficients were tabulated for each variable and compared between scoring pairs (A-M1, A-M2, M1-M2). The 26,876 epochs (224 hours of recording time) were analyzed. For sleep staging, agreement/kappa were A-M1: 78%/0.67, A-M2: 73%/0.61, and M1-M2: 82%/0.73. The mean respiratory disturbance indexes were M1: 20.6+/-23.0, M2: 22.5+/-24.5, and A: 23.7+/-23.4 events per hour of sleep. The respiratory disturbance index concordance between each scoring pair was excellent (intraclass correlation coefficients > or = 0.95 for all pairs), although there was disagreement in the classification of moderate sleep-disordered breathing (percentage of positive agreement: A-M1, 37.5% and A-M2, 44.4%) defined as a respiratory disturbance index between 15 and 30 events per hour of sleep. For respiratory-event detection, agreement/kappa were A-M1 and A-M2: 90%/0.66 and M1-M2: 95%/0.82. The agreement and kappa for limb movement detection were A-M1: 93%/0.68, A-M2: 92%/0.66, and M1-M2: 96%/0.77. The scoring of arousals was less reliable (agreement range: 76%-84%, kappa range: 0.28-0.57) for all pairs. CONCLUSIONS: Agreement between manual scorers in a population with moderate sleep-disordered breathing was close to the average pairwise agreement of 87% reported in the Sleep Heart Health Study. The automated classification of sleep stages was also close to this standard. The automated scoring system holds promise as a rapid method to score polysomnographic records, but expert verification of the automated scoring is required.
STUDY OBJECTIVES: To assess the accuracy of an automated system (Morpheus I Sleep Scoring System) for analyzing and quantifying polysomnographic data from a population with sleep-disordered breathing. SETTING: Sleep laboratory affiliated with a tertiary care academic medical center. MEASUREMENTS AND RESULTS: 31 diagnostic polysomnograms were blindly analyzed prospectively with the investigational automated system and manually by 2 registered polysomnography technologists (M1 & M2) from the same laboratory. Sleep stages, arousals, periodic limb movements, and respiratory events (apneas and hypopneas) were scored by all 3. Agreement, Cohen kappa, and intraclass correlation coefficients were tabulated for each variable and compared between scoring pairs (A-M1, A-M2, M1-M2). The 26,876 epochs (224 hours of recording time) were analyzed. For sleep staging, agreement/kappa were A-M1: 78%/0.67, A-M2: 73%/0.61, and M1-M2: 82%/0.73. The mean respiratory disturbance indexes were M1: 20.6+/-23.0, M2: 22.5+/-24.5, and A: 23.7+/-23.4 events per hour of sleep. The respiratory disturbance index concordance between each scoring pair was excellent (intraclass correlation coefficients > or = 0.95 for all pairs), although there was disagreement in the classification of moderate sleep-disordered breathing (percentage of positive agreement: A-M1, 37.5% and A-M2, 44.4%) defined as a respiratory disturbance index between 15 and 30 events per hour of sleep. For respiratory-event detection, agreement/kappa were A-M1 and A-M2: 90%/0.66 and M1-M2: 95%/0.82. The agreement and kappa for limb movement detection were A-M1: 93%/0.68, A-M2: 92%/0.66, and M1-M2: 96%/0.77. The scoring of arousals was less reliable (agreement range: 76%-84%, kappa range: 0.28-0.57) for all pairs. CONCLUSIONS: Agreement between manual scorers in a population with moderate sleep-disordered breathing was close to the average pairwise agreement of 87% reported in the Sleep Heart Health Study. The automated classification of sleep stages was also close to this standard. The automated scoring system holds promise as a rapid method to score polysomnographic records, but expert verification of the automated scoring is required.
Authors: Timothy I Morgenthaler; Janet B Croft; Leslie C Dort; Lauren D Loeding; Janet M Mullington; Sherene M Thomas Journal: J Clin Sleep Med Date: 2015-09-15 Impact factor: 4.062
Authors: Lien Desteghe; Jeroen M L Hendriks; R Doug McEvoy; Ching Li Chai-Coetzer; Paul Dendale; Prashanthan Sanders; Hein Heidbuchel; Dominik Linz Journal: Clin Res Cardiol Date: 2018-04-12 Impact factor: 5.460
Authors: Sabrina L Wendt; Peter Welinder; Helge B D Sorensen; Paul E Peppard; Poul Jennum; Pietro Perona; Emmanuel Mignot; Simon C Warby Journal: Clin Neurophysiol Date: 2014-11-10 Impact factor: 3.708
Authors: Atul Malhotra; Magdy Younes; Samuel T Kuna; Ruth Benca; Clete A Kushida; James Walsh; Alexandra Hanlon; Bethany Staley; Allan I Pack; Grace W Pien Journal: Sleep Date: 2013-04-01 Impact factor: 5.849
Authors: Cathy A Goldstein; Richard B Berry; David T Kent; David A Kristo; Azizi A Seixas; Susan Redline; M Brandon Westover Journal: J Clin Sleep Med Date: 2020-04-15 Impact factor: 4.062