| Literature DB >> 31240415 |
Jonas Denck1,2,3, Wilfried Landschütz4, Knud Nairz5, Johannes T Heverhagen5, Andreas Maier6, Eva Rothgang7.
Abstract
Although the level of digitalization and automation steadily increases in radiology, billing coding for magnetic resonance imaging (MRI) exams in the radiology department is still based on manual input from the technologist. After the exam completion, the technologist enters the corresponding exam codes that are associated with billing codes in the radiology information system. Moreover, additional billing codes are added or removed, depending on the performed procedure. This workflow is time-consuming and we showed that billing codes reported by the technologists contain errors. The coding workflow can benefit from an automated system, and thus a prediction model for automated assignment of billing codes for MRI exams based on MRI log data is developed in this work. To the best of our knowledge, it is the first attempt to focus on the prediction of billing codes from modality log data. MRI log data provide a variety of information, including the set of executed MR sequences, MR scanner table movements, and given a contrast medium. MR sequence names are standardized using a heuristic approach and incorporated into the features for the prediction. The prediction model is trained on 9754 MRI exams and tested on 1 month of log data (423 MRI exams) from two MRI scanners of the radiology site for the Swiss medical tariffication system Tarmed. The developed model, an ensemble of classifier chains with multilayer perceptron as a base classifier, predicts medical billing codes for MRI exams with a micro-averaged F1-score of 97.8% (recall 98.1%, precision 97.5%). Manual coding reaches a micro-averaged F1-score of 98.1% (recall 97.4%, precision 98.8%). Thus, the performance of automated coding is close to human performance. Integrated into the clinical environment, this work has the potential to free the technologist from a non-value adding an administrative task, therefore enhance the MRI workflow, and prevent coding errors.Entities:
Keywords: Machine learning; Magnetic resonance imaging; Medical coding; Reimbursement; Workflow enhancement
Mesh:
Year: 2019 PMID: 31240415 PMCID: PMC6841869 DOI: 10.1007/s10278-019-00241-z
Source DB: PubMed Journal: J Digit Imaging ISSN: 0897-1889 Impact factor: 4.056
Fig. 1Time line of an MRI exam. All logs from the subject registration on the MRI host computer to the end of the last sequence describe a single instance
Examples of original and standardized sequence name pairs
| Original sequence name | Standardized sequence name |
|---|---|
| ep2d_diff_3b_Abdomen | ep2d_diff tra |
| ep2d_diff_b50_300_800_tra_4Scan_p3 | ep2d_diff tra |
| t2_haste_fs_tra_3mm_mbh | t2 haste fs tra bh |
| t1_vibe_fs_tra_caipi_15 min | t1 vibe fs tra bh |
Abbreviations: ep2d = two-dimensional (2d) echo-planar imaging; diff, diffusion; tra, transverse; mbh, multiple breath hold
The first two sequences listed in the table represent the same diffusion weighted sequence and are therefore mapped to the same standardized sequence name
See http://www.revisemri.com/questions/misc/mri_abbrev for an extensive list of common MRI abbreviations that explain the remaining sequence terms
Fig. 2Distribution of the number of procedure codes per MRI exam in the training dataset. If no procedure code was reported for an MRI exam, it was aborted before images sufficient for diagnosing were acquired
Evaluation scores and label metrics for the test prediction using different classification methods
| Binary relevance | Ensemble of classifier chains | Ensemble of classifier chains | |||
|---|---|---|---|---|---|
| MLP | SVM RBF | Random forest | MLP | Random forest | |
| Micro F1-score | 97.5% | 95.3% | 97.3% | 97.8% | 97.5% |
| Micro precision | 97.4% | 96.4% | 97.7% | 97.5% | 97.5% |
| Micro recall | 97.6% | 94.3% | 97.0% | 98.1% | 97.5% |
| Subset accuracy | 90.8% | 84.6% | 90.6% | 91.7% | 91.3% |
| Label cardinality | 4.1 | 4.0 | 4.1 | 4.1 | 4.1 |
| Label density | 0.2 | 0.2 | 0.2 | 0.2 | 0.2 |
| Label diversity | 57 | 40 | 47 | 55 | 55 |
Fig. 3The Venn diagram illustrates the performance difference of automated and manual coding with regard to the subset accuracy. Here, the ECC-MLP model was used for automated coding. In a significant share of the test instances, automated coding was superior to manual coding (5.6%)