Literature DB >> 24347844

Voice recognition is here comma like it or not period.

Matthew A Fox1, Carl J Aschkenasi1, Arjun Kalyanpur1.   

Abstract

Voice recognition (VR) technology needs improvement, but is as integral to the current practice of radiology as Radiology Information Systems and Picture Archival and Communication Systems. In the 1990s, the radiology community gave VR technology a rather lukewarm reception, but since then it has emerged as the predominant method of radiology reporting in the United States. In this article, we examine how VR technology works, outline the positive and negative aspects of VR technology on work flow, identify common VR transcription errors and review the discussion on VR adoption in the recent literature. We add to the discussion our personal experiences in an international teleradiology group.

Entities:  

Keywords:  Acoustic; language; phoneme; transcription; voice recognition

Year:  2013        PMID: 24347844      PMCID: PMC3843322          DOI: 10.4103/0971-3026.120252

Source DB:  PubMed          Journal:  Indian J Radiol Imaging        ISSN: 0970-2016


Introduction

Over the past decade major advances in technology have not only impacted the way we image our patients, but also fundamentally changed the workflow in the practice of Radiology. With the emergence of Radiology Information Systems (RIS) and Picture Archival and Communication Systems (PACS), the process of reviewing radiology exams migrated from a film and paper based process to digitized data and media. A key component which has contributed to further streamlining and digitizing the radiology work flow has been the integration of Voice Recognition (VR) technology into the process of radiology report production. To better understand the role VR is playing in the evolving radiology workflow, we begin with a review of how VR technology works.

VR in a Nutshell

Voice recognition (VR) technology has made tremendous strides since its inception, attributed to IBM, which showcased a prototype of the technology at the 1962 Seattle World's Fair.[1] The “IBM shoebox” was a rudimentary computer that could perform speech recognition of 16 spoken words and the digits 0 through 9, producing either a printed output of simple arithmetic computations or serial activation of light bulbs. The basic elements of that prototype still serve modern-day VR systems. A microphone and sound card are used as an analog-to-digital-converter (ADC) of the continuous spoken sound waves that are transformed into vibrations which are then converted into digital data. The digital data is analyzed by the VR software, which breaks it down to the smallest sound segments used in speech, called phonemes. The VR software then runs the phoneme segmented digital data through a series of statistical algorithms in order to identify and contextualize strings of recorded phonemes into words and phrases found in its language library. In the last step, the VR software then outputs the digital data as a coherent text or generates a computer command [Figure 1].
Figure 1

VR step model

VR step model Most VR software research and development is focused on the penultimate step in this process, phoneme contextualization. Two key components of this step are the acoustic model algorithms and the language model algorithm.[2] The acoustic model is composed of a library of word phoneme compositions, including common variations in pronunciation, which are used to statistically predict the most probable word spoken. The language model is composed of a library of domain-specific words and phrases used to statistically predict the most likely word sequence spoken. Generally, the acoustic model performs the first calculation of the expected word which is then checked for contextualization based on statistical probability calculated by the language model. This integrated model assures a higher degree of accuracy. For example, if the acoustic model predicts that the most likely words spoken are “No intracranial pass or hemorrhage,” then the combined probabilities of both the acoustic and language model would be “No intracranial mass or hemorrhage.

Adoption of VR Technology in Radiology

A recent survey of over 1000 physicians from all fields of medicine revealed that over 75% use VR technology in their daily practice.[3] A similar informal poll in Diagnostic Imaging (Sept. 2011) yielded 80% of respondents as users, but 30% reporting dissatisfaction with the technology. Financial pressures and interests, most notably in the United States, have incentivized automated reporting rather than more costly human transcriptionists. Similar pressures have also meant a push for greater patient turnover in all clinical settings, especially in the emergency room (ER). VR has been promoted as a way to shorten the gap between scan time and report arrival. VR also dovetails with the digital migration of healthcare to electronic medical records (EMRs), and filmless technology supported by Radiology Information Systems (RIS) and Picture Archival and Communication Systems (PACS). Furthermore, the burgeoning practice of structured medical reporting also parallels the increased adoption of VR, where templates and macros can be used (for better or worse) to create modular reports for many commonly encountered diagnoses.[4] VR technology also provides an effective platform for promoting standardized reporting methods and terminology. In addition, the affordability of high-powered processors, as well as cheap memory, permits the use of VR technology with off-the-shelf computers and workstations. But VR has always had its detractors, and still does, despite it being the heir apparent to traditional human transcriptionists. At the institution where one author (CJA) trained, there was a marked schism between those for and against the adoption of VR, and interestingly it was not predicated along the lines of older recalcitrant physicians versus their younger colleagues. Some doctors expressed a concern for the soon-to-be laid-off or “re-employed” transcriptionists. Others were doubtful about whether there would be a net gain in productivity when the chief activities of the transcriptionist were to be taken up by the radiologist. Many were concerned about errors, misappropriations, etc., that would be overlooked by a rushed radiologist, but could be caught in many instances by experienced transcriptionists.

Terror-Prune Retorting (Error-Prone Reporting)

The leading drawback for most radiologists using VR is that it can make many errors, some of which can be difficult to detect when the speaker himself proofs the report. Common voice recognition errors [Table 1] cause more than just frustration. These errors have potential medicolegal consequences and have real financial consequences as well. It has been estimated that the time spent correcting errors by VR-generated reports is worth about $80,000 dollars per radiologist, annually.[5] Reported VR transcription error rates vary from as low as 2% to as high as 90% prior to proofreading. For example, a British study in 2008 showed a statistically significant twofold difference in error rate between dictation-transcription (DT) and VR; however, the overall error rate was quite low at 3.8%.[6] This is vastly different from a report from Brown University which showed an error rate of 89% for VR compared to 10% for DT.[5] Interestingly, even after proofing by the radiologist, the VR error rate was still very high at 35%.
Table 1

Common voice recognition errors

Common voice recognition errors These numbers, however, are difficult to interpret as they reflect different VR software implemented and assessed in different ways along the radiology workflow. However, several types of errors are frequently encountered, including wrong-word substitution, nonsense phrases, and missing words.[7] Common examples of substitution of a wrong word were sound-alike words, such as “normally” substituted for “minimally,” “ill-defined” substituted for “well-defined,” or “location” substituted for “dislocation.” Nonsense phrases may arise simply due to the VR software error in parsing of the phonemes. For example, the word “retroactive” would be incorrectly transcribed into “rich or active” or “valuable” into “buy you a bull.” Another group of common errors involves numbers, measurements, and dates, which appear to be prone to VR transcription errors, and are difficult to catch when proofreading reports. One author (MAF) learned that a combination of improving the microphone hardware quality (i.e., USB versus audio jack), performing additional software VR training modules, as well as specific word training of the VR software led to an appreciable improvement in the VR transcription accuracy. Others have turned to an increased use of templates, reducing the words which are actually transcribed.

Impact on Training

Dr. Sanjeev Bhalla, chief of Thoracic Imaging and Assistant Program Director of the residency program at Mallinckrodt Institute of Radiology/Washington University in St Louis, raises further concerns regarding VR from a resident training standpoint.[8] Dr. Bhalla is concerned (and he is not alone) that VR diverts residents’ attention from detailed review of the images, the nidus of their education, to a detailed review of the text in the dictation window. The time spent editing reports results in fewer cases covered and less training experience. He also observes that in a medical environment where the report is quickly available, direct consultations between radiologist and referring clinician decrease in both quantity and quality. When such consultations do occur, they are more apt to occur by phone, and revolve around the phrasing of a report rather than a personal exchange at the monitor in which imaging findings and clinical findings are compared.[9] The benefit of such direct discussion is better understanding of the clinical problem, its imaging manifestations, and ultimately a more meaningful final report reached through consensus impression. The learning potential of such interactions is enormous and may be jeopardized in a VR-mediated communication. Of course, among the counterarguments to these concerns would be the fact that as medical communicators, the report is our product, thus it is important to teach trainees to craft a cogent report with the tools that they will most likely use after they complete their training. A study conducted at Harvard Medical School concluded that in order for speech recognition to prove beneficial to radiology resident education, it must include adequate dedicated technical training and support including the use of specialized functions such as customizable templates/macros.[10] With improved user proficiency, the residents and attending physicians then need to be cognizant of properly incorporating the technology into the preview, review, and dictation process revolving image interpretation without losing sight of the necessary detailed visualization of the imaging findings by the residents.

The Need for Speed

Drawbacks notwithstanding, it is hard to do better than instantaneous, and that is chiefly the metric in which VR excels. Numerous studies have shown improved report turnaround time (TAT) with VR.[1112] This is true for multiple languages and important particularly for our international practice, for English spoken by users whose first language is not English.[13] Rapid TAT is not only a clinically important issue in some areas, particularly in the ER environment, but also has become a critical component in cost containment in the ER, where rapid patient throughput is highly valued. In addition, continuity of care is significantly enhanced by the decreased report TAT, allowing for patient transfer with available imaging reports to level I trauma centers or other tertiary care facilities. Finally, it could further be surmised that rapid TATs may reduce risk in a litigious medicolegal environment.

Our Experience

The authors have worked with traditional human transcription systems, VR-based workflow, and a combination of these, where transcriptionists proofread VR reports. Furthermore, we were in training or early professional practice during the transition to widespread VR use. We have learned that radiologists must slow down and articulate words clearly to work effectively with VR despite “training” functions included in sophisticated VR software. Human transcriptionists can deal with rushed, run-on, or slurred dictation styles, which VR simply cannot. Although the software's adaptive technology learns from the user, this technology has its limitations. Even the most advanced user will find it necessary to compromise or work around some small inherent foible of the software in order to maximize its advantages. In addition, we have learned to appreciate the use of templates in reporting, a convenience which saves time during routine cases, enabling us to devote more attention to complex cases requiring carefully considered language. In order to leverage the advantages of VR over its disadvantages [Table 2] we have implemented a two-tiered approach in our practice with VR as the “front-end” reporting modality, followed by human editing. The medical transcriptionist still adds value in that he can clean up errors missed by both report author and VR software and ensure that exams are properly identified with the relevant patients and that specific formatting requests of a client are met. For the demands of our practice, which covers numerous modalities for globally dispersed clients with different needs, expectations, and medical cultures, and globally dispersed radiologists with differing approaches to practice and differing accents, we have found VR an indispensable tool (but not without its challenges), which helps unify our group with a common reporting interface.
Table 2

Pros and cons of using voice recognition

Pros and cons of using voice recognition
  10 in total

1.  Voice recognition software: effect on radiology report turnaround time at an academic medical center.

Authors:  Arun Krishnaraj; Joseph K T Lee; Sandra A Laws; T Jay Crawford
Journal:  AJR Am J Roentgenol       Date:  2010-07       Impact factor: 3.959

2.  Six characteristics of effective structured reporting and the inevitable integration with speech recognition.

Authors:  David Liu; Mark Zucherman; William B Tulloss
Journal:  J Digit Imaging       Date:  2006-03       Impact factor: 4.056

3.  Impact of PACS and voice-recognition reporting on the education of radiology residents.

Authors:  Antonio J Gutierrez; Mark E Mullins; Robert A Novelline
Journal:  J Digit Imaging       Date:  2005-06       Impact factor: 4.056

4.  Voice recognition dictation: radiologist as transcriptionist.

Authors:  John A Pezzullo; Glenn A Tung; Jeffrey M Rogg; Lawrence M Davis; Jeffrey M Brody; William W Mayo-Smith
Journal:  J Digit Imaging       Date:  2008-12       Impact factor: 4.056

5.  Frequency and spectrum of errors in final radiology reports generated with automatic speech recognition technology.

Authors:  Leslie E Quint; Douglas J Quint; James D Myles
Journal:  J Am Coll Radiol       Date:  2008-12       Impact factor: 5.532

6.  The effect of voice recognition software on comparative error rates in radiology reports.

Authors:  S McGurk; K Brauer; T V Macfarlane; K A Duncan
Journal:  Br J Radiol       Date:  2008-07-15       Impact factor: 3.039

7.  Improvement of report workflow and productivity using speech recognition--a follow-up study.

Authors:  Tomi Kauppinen; Mika P Koivikko; Juhani Ahovuo
Journal:  J Digit Imaging       Date:  2008-04-24       Impact factor: 4.056

8.  Impact of a voice recognition system on radiology report turnaround time: experience from a non-english-speaking South Asian country.

Authors:  Waseem Akhtar; Arif Ali; Kashif Mirza
Journal:  AJR Am J Roentgenol       Date:  2011-04       Impact factor: 3.959

9.  Voice recognition in radiology reporting.

Authors:  L H Schwartz; P Kijewski; H Hertogen; P S Roossin; R A Castellino
Journal:  AJR Am J Roentgenol       Date:  1997-07       Impact factor: 3.959

10.  The pros and cons of implementing PACS and speech recognition systems.

Authors:  D B Hayt; S Alexander
Journal:  J Digit Imaging       Date:  2001-09       Impact factor: 4.056

  10 in total
  2 in total

1.  Frequency and analysis of non-clinical errors made in radiology reports using the National Integrated Medical Imaging System voice recognition dictation software.

Authors:  R E Motyer; S Liddy; W C Torreggiani; O Buckley
Journal:  Ir J Med Sci       Date:  2016-10-01       Impact factor: 1.568

2.  Limitations in and Solutions for Improving the Functionality of Picture Archiving and Communication System: an Exploratory Study of PACS Professionals' Perspectives.

Authors:  Mona Alhajeri; Syed Ghulam Sarwar Shah
Journal:  J Digit Imaging       Date:  2019-02       Impact factor: 4.056

  2 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.