| Literature DB >> 25351845 |
Maree Johnson1, Samuel Lapkin, Vanessa Long, Paula Sanchez, Hanna Suominen, Jim Basilakis, Linda Dawson.
Abstract
BACKGROUND: To undertake a systematic review of existing literature relating to speech recognition technology and its application within health care.Entities:
Mesh:
Year: 2014 PMID: 25351845 PMCID: PMC4283090 DOI: 10.1186/1472-6947-14-94
Source DB: PubMed Journal: BMC Med Inform Decis Mak ISSN: 1472-6947 Impact factor: 2.796
Search strategies OVID Embase, Medline, PreMedline
|
| ||
| 1 | automatic speech recognition/ | 469 |
| 2 | ((voice or speech) adj (recogni* or respon*)).tw. | 2516 |
| 3 | or/1-2 | 27490 |
| 4 | exp research/ | 380483 |
| 5 | (qualitative* or quantitative* or mixed method* or descriptive* or research*).tw. | 1194784 |
| 6 | or/4-5 | 14148120 |
| 7 | 3 and 6 | 483 |
| 8 | limit 7 to yr = “2000 -Current” | 433 |
|
| ||
| 1 | Speech Recognition Software | 416 |
| 2 | ((voice or speech) adj (recogni* or respon*)).tw. | 2081 |
| 3 | or/1-2 | 2263 |
| 4 | exp Research/ | 224487 |
| 5 | (qualitative* or quantitative* or mixed method* or descriptive* or research*).tw. | 840821 |
| 6 | or/4-5 | 971456 |
| 7 | 3 and 6 | 360 |
| 8 | limit 7 to yr = “2000 -Current” | 319 |
|
| ||
| 1 | ((voice or speech) adj (recogni* or respon*)).tw. | 140 |
| 2 | (qualitative* or quantitative* or mixed method* or descriptive* or research*).tw. | 94513 |
| 3 | 1 and 2 | 20 |
| 4 | limit 3 to yr = “2000 -Current” | 19 |
Note that Speech Recognition Software refers to a MeSH term. * = wildcard.
Figure 1Selection of studies for the review.
SR Quality scoring of included studies - Mixed Methods Appraisal Tool (MMAT)-Version 2011
| Al-Aynati 2003
[ | Alapetite, 2008
[ | Alapetite, 2009
[ | Callaway, 2002
[ | Derman, 2010
[ | Devine, 2000
[ | Irwin, 2007
[ | Kanal, 2001
[ | Koivikko, 2008
[ | Langer, 2002
[ | Mohr, 2003
[ | NSLHD 2012
[ | Singh, 2011
[ | Zick, 2001
[ | |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
|
| ||||||||||||||
| Clear research questions | Yes | Yes | Yes | Yes | Yes | Yes | Yes | Yes | Yes | Yes | Yes | Yes | Yes | Yes |
| Appropriate data collected | Yes | Yes | Yes | Yes | Yes | Yes | Yes | Yes | Yes | Yes | Yes | Yes | Yes | Yes |
|
| ||||||||||||||
| Appropriate qualitative data sources | ||||||||||||||
| Appropriate qualitative method | ||||||||||||||
| Description of the context | ||||||||||||||
| Discussion of researchers’ reflexivity | ||||||||||||||
|
| ||||||||||||||
| Appropriate randomization | Yes | No | ||||||||||||
| Allocation concealment and/or blinding | Yes | No | ||||||||||||
| Complete outcome data | Yes | Yes | ||||||||||||
| Low withdrawal/drop out | Yes | Yes | ||||||||||||
|
| ||||||||||||||
|
| ||||||||||||||
| Recruitment minimized bias | No | |||||||||||||
| Appropriate outcome measures | Yes | |||||||||||||
| Intervention & control group comparable | Yes | |||||||||||||
| Complete outcome data/acceptable response rate | Yes | |||||||||||||
|
| ||||||||||||||
| Appropriate sampling1 | No | Yes | Yes | No | Yes | Yes | No | Yes | Yes | Yes | No | |||
| Appropriate sample2 | No | Yes | Yes | Yes | Yes | Yes | Yes | Yes | Yes | Yes | Yes | |||
| Appropriate measurement (valid/standard) | Yes | Yes | Yes | Yes | Yes | Yes | Yes | Yes | Yes | Yes | Yes | |||
| Acceptable response rate | Yes | Yes | Yes | Yes | Yes | No | Yes | Yes | No | Yes | Yes | |||
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1Sampling strategy relevant to address the quantitative research question. Consider whether the source of sample is relevant to the population under study; when appropriate, there is a standard procedure for sampling; the sample size is justified (using power calculation for example).
2Sample representative of the population under study. Consider whether inclusion and exclusion criteria are explained; reasons why certain eligible individuals chose not to participate are explained.
3Scores ranged from 0–6.
Summary of speech recognition (SR) review results
| Author | Aim | Setting | Outcome measures | Results |
|---|---|---|---|---|
| Year | Sample | |||
| Country Design | Speech technology (ST) | |||
| Design | ||||
| Al-Aynati and Chorneyko 2003 [ | To compare SR software with HT for generating pathology reports |
| 1. Accuracy rate |
|
|
| 2. Recognition/ Transcription errors | SR: 93.6 HT: 99.6 | ||
| Canada Experimental |
|
| ||
| SR: 6.7 HT: 0.4 | ||||
| Mohr et al. 2003 [ | To compare SR software with HT for clinical notes |
| 1. Dictation/recording time + transcription (minutes) = Report Turnaround Time (RTT). |
|
|
| ||||
| SR: (Recording + transcription) = 23.7 | ||||
| HT: (Dictation + transcription) = 25.4 | ||||
| USA Experimental |
| |||
|
| SR: 87.3% (CI 83.3, 92.3) productive compared to HT. | |||
|
| ||||
| SR: (Recording + transcription) = 65.2 | ||||
| HT: (Dictation + transcription) = 38.1 | ||||
| SR: 63.3% (CI 54.0, 74.0) productive compared to HT. | ||||
|
| ||||
| SR: (Recording + transcription) = 36.5 | ||||
| HT: (Dictation + transcription) = 30.5 | ||||
| SR: 55.8% (CI 44.6, 68.0) productive compared to HT. | ||||
| Author, secretary, type of notes were predictors of productivity (p < 0.05). | ||||
| NSLHD 2012 [ | To compare accuracy and time between SR software and HT to produce emergency department reports |
| 1. RTT |
|
| Australian Experimental |
| SR: 1.07 (46 sec, 1.32) | ||
|
| HT: 3.32 (2.45, 4.35) | |||
| HT: Spelling and punctuation errors | ||||
| SR: Occasional misplaced words | ||||
| Alapetite, 2008 [ | To evaluate the impact of background |
| 1. Word Recognition Rate (WRR) |
|
| Denmark Non-experimental | noise (sounds of alarms, aspiration, metal, people talking, scratch, silence, ventilators) and other factors affecting SR accuracy when used in operating rooms |
|
| |
| Microphone 1: Headset 83.2% | ||||
|
| Microphone 2: Handset 73.9% | |||
|
| ||||
| Command 81.6% | ||||
| Free text 77.1% | ||||
|
| ||||
| Scratch 66.4% | ||||
| Silence 86.8% | ||||
|
| ||||
| Male 76.8% | ||||
| Female 80.3% | ||||
| Alapetite et al. 2009 [ | To identify physician’s perceptions, attitudes and expectations of SR technology. |
| 1. Users’ expectation and experience |
|
| Denmark Non-experimental |
| Predominant response noted. |
| |
|
| ||||
|
| ||||
|
| ||||
|
| ||||
|
| ||||
|
| ||||
|
| ||||
|
| ||||
|
| ||||
|
| ||||
| Callaway et al. 2002 [ | To compare an off the shelf SR software with manual transcription services for radiology reports |
| 1. RTT (referred to as TAT) |
|
| USA Non-experimental |
| 2. Costs |
| |
| Facility 2: 26600 reports |
| |||
| Facility 3: 5109 reports |
| |||
|
|
| |||
| Professional 4.0 |
| |||
|
| ||||
| Derman et al. 2010 [ | To compare SR with existing methods of data entry for the creation of electronic progress notes |
| 1. Perceived usability |
|
| Canada Non-experimental |
| 2. Perceived time savings | 50% prefer SR | |
| 3. Perceived impact |
| |||
|
| ||||
|
| ||||
|
| ||||
|
| ||||
| Devine et al. 2000 [ | To compare ‘out-of-box’ performance of 3 continuous SR software packages for the generation of medical reports. |
| 1. Recognition errors (mean error rate) |
|
| USA Non-experimental |
| 2. Dictation time |
| |
| 3. Completion time |
| |||
|
| 4. Ranking |
| ||
|
| ||||
|
| 5. Preference |
| ||
|
| ||||
|
| ||||
| Irwin et al. 2007 [ | To compare SR features and functionality of 4 dental software application systems. |
| 1. Training time |
|
| USA Non-experimental |
| 2. Charting time |
| |
| 3. Completion | ||||
|
| 4. Ranking |
| ||
|
|
| |||
|
|
| |||
|
| ||||
| Kanal et al. 2001 [ | To determine the accuracy of continuous SR for transcribing radiology reports |
| 1. Error rates |
|
| USA Non-experimental |
|
| ||
|
| ||||
|
|
| |||
| Koivikko et al. 2008 [ | To evaluate the effect of speech recognition onadiology workflow systems over a period of 2 years |
| 1. RTT (referred to as TAT) at 3 collection points: |
|
| Finland Non-experimental |
| HT: 2005 (n = 6037) | HT: 1486 ± 4591 | |
|
| SR1: 2006 (n = 6486) | SR | ||
| Recognition System (Philips Electronics) | SR2: 2007 (n = 9072) | SR | ||
| HT: cassette-based reporting | 2. Reports completed ≤ 1 hour |
| ||
| SR1: SR in 2006 | HT: 26 | |||
| SR2: SR in 2007 | SR | |||
|
| ||||
| 10-15 minutes training in SR | ||||
| Langer 2002 [ | To compare impact of SR on radiologist productivity. Comparison of 4 workflow systems |
| 1. RTT (referred to as TAT) |
|
| USA Non-experimental |
| 2. Report productivity (RP), number of reports per day |
| |
|
| RTT: 48.2 ± 50 RP: 240 | |||
|
|
| |||
|
| RTT: 15.5 ± 93 RP: 311 | |||
|
| ||||
|
| RTT: 13.3 ± 119 (t value at 10%) RP: 248 | |||
|
| ||||
| RTT: 15.7 ± 98 (t value at 10%) RP: 310 | ||||
| Singh et al. 2011 [ | To compare accuracy and turnaround |
| 1. RTT (referred to as TAT) |
|
| USA Non-experimental | times between SR software and traditional transcription service (TS) when used for generating surgical pathology reports |
| 2. Reports completed ≤ 1 day |
|
|
| 3. Reports completed ≤ 2 day |
| ||
|
|
| |||
|
|
| |||
| Phase 0: 22 | ||||
|
| Phase 1: 24 | |||
| Phase 2–4: 36 | ||||
|
| ||||
| Phase 0: 54 | ||||
| Phase 1: 60 | ||||
| Phase 2–4: 67 | ||||
| Zick et al. 2001 [ | To compare accuracy and RTT between |
| 1. RTT (referred to as TAT) |
|
| USA Non-experimental | SR software and traditional transcription service (TS) when used for recording in patients’ charts in ED |
| 2. Accuracy | SR: 3.55 TS: 39.6 |
| 3. Errors per chart |
| |||
|
| 4. Dictation and editing time | SR: 98.5 (98.2-98.9) TS: 99.7 (99.6-99.8) | ||
| 4. Throughput |
| |||
| SR: 2.5 (2–3) TS: 1.2 (0.9-1.5) | ||||
|
| ||||
| SR: 3.65 (3.35-3.95) TS: 3.77 (3.43-4.10) | ||||
|
| ||||
| SR: 54.5 (49.6-59.4) TS: 14.1 (11.1-17.2) |
Report productivity (RP): Normalises the output of staff to the daily report volume.
Note: SR = speech recognition ST = speech technology HT = human transcription RTT = report turnaround time WRR = word recognition rate PACS = picture archiving and communication system RP = report productivity TS = traditional transcription service ED = emergency department Sig. = Significant Diff = difference. TAT = turnaround time, equivalent to RTT.