| Literature DB >> 30357017 |
Xinyu Li1, Yanyi Zhang1, Mengzhu Li1, Shuhong Chen1, Farneth R Austin2, Ivan Marsic1, Randall S Burd2.
Abstract
We present a multimodal deep-learning structure that automatically predicts phases of the trauma resuscitation process in real-time. The system first pre-processes the audio and video streams captured by a Kinect's built-in microphone array and depth sensor. A multimodal deep learning structure then extracts video and audio features, which are later combined through a "slow fusion" model. The final decision is then made from the combined features through a modified softmax classification layer. The model was trained on 20 trauma resuscitation cases (>13 hours), and was tested on 5 other cases. Our results showed over 80% online detection accuracy with 0.7 F-Score, outperforming previous systems.Entities:
Keywords: Activity recognition; Kinect; deep learning; multimodal sensing; process phase recognition; trauma resuscitation
Year: 2016 PMID: 30357017 PMCID: PMC6196033 DOI: 10.1109/UEMCON.2016.7777912
Source DB: PubMed Journal: Ubiquitous Comput Electron Mob Commun Conf (UEMCON) IEEE Annu