| Literature DB >> 34870242 |
Roshan P Mathews1, Mahesh Raveendranatha Panicker1, Abhilash R Hareendranathan2.
Abstract
The COVID-19 pandemic has accelerated the need for automatic triaging and summarization of ultrasound videos for fast access to pathologically relevant information in the Emergency Department and lowering resource requirements for telemedicine. In this work, a PyTorch based unsupervised reinforcement learning methodology which incorporates multi feature fusion to output classification labels, segmentation maps and summary videos for lung ultrasound is presented. The use of unsupervised training eliminates tedious manual labeling of key-frames by clinicians opening new frontiers in scalability in training using unlabeled or weakly labeled data. Our approach was benchmarked against expert clinicians from different geographies displaying superior Precision and F1 scores (over 80% and 44%).Entities:
Keywords: Attention ensemble encoders; PyTorch; Python; Ultrasound; Unsupervised reinforcement learning; Video summarization
Year: 2021 PMID: 34870242 PMCID: PMC8628609 DOI: 10.1016/j.simpa.2021.100185
Source DB: PubMed Journal: Softw Impacts ISSN: 2665-9638
Fig. 1Graphical abstract of the video summarization algorithm. The input video frames are encoded using standard encoder architectures and then summarized using an LSTM decoder. The summarized videos contain pathologically relevant key frames and are overlaid with classification scores and segmentation maps for easy diagnosis by clinicians.
Quantitative Metrics for the video summarization methodology.
| Approach | Encoder - LSTM Decoder | Precision | Recall | ReF | |
|---|---|---|---|---|---|
| Classification | 46.38 | 16.11 | 23.91 | 78.62 | |
| General | Segmentation | 61.58 | 24.87 | 35.43 | 75.08 |
| Auto-Encoding | 59.46 | 22.84 | 33.00 | 78.11 | |
Fig. 2Preferential (diagnostically important) summarization aspect of the algorithm. (A) Presence of ambiguous regions in the ultrasound scan and (B) the preferential selection of pathologically relevant structures like the B-lines for the summary.
Quantitative Metrics for ablation studies involving rewards in parts.
| Approach | Rewards | Encoder-LSTM Decoder | Precision | Recall | |
|---|---|---|---|---|---|
| Classification | 46.93 | 16.26 | 24.15 | ||
| General | rep + div | Segmentation | 60.78 | 24.20 | 34.62 |
| Auto-Encoding | 58.19 | 22.30 | 32.24 | ||
Fig. 3Web-application deployment. The tool has the option to select the classification labels and the segmentation maps to be overlaid on the frames. Provision is also given to save the frame features. The tool will show the summary video, a collage of random frames from the summary video and the frame numbers.
| Current code version | v1 |
| Permanent link to code/repository used for this code version | |
| Permanent link to Reproducible Capsule | |
| Legal Code License | MIT License |
| Code versioning system used | git |
| Software code languages, tools, and services used | python (Anaconda distribution), Spyder (IDE) |
| Compilation requirements, operating environments & dependencies | python 3.8, PyTorch 1.10, segmentation_models_pytorch, torchvision, albumentations, opencv, h5py, numpy, matplotlib, shutil, tkinter, streamlit. |
| If available Link to developer documentation/manual | |
| Support email for questions |