| Literature DB >> 34127879 |
Thanassis Mavropoulos1, Spyridon Symeonidis1, Athina Tsanousa1, Panagiotis Giannakeris1, Maria Rousi1, Eleni Kamateri1, Georgios Meditskos1, Konstantinos Ioannidis1, Stefanos Vrochidis1, Ioannis Kompatsiaris1.
Abstract
The details presented in this article revolve around a sophisticated monitoring framework equipped with knowledge representation and computer vision capabilities, that aims to provide innovative solutions and support services in the healthcare sector, with a focus on clinical and non-clinical rehabilitation and care environments for people with mobility problems. In contemporary pervasive systems most modern virtual agents have specific reactions when interacting with humans and usually lack extended dialogue and cognitive competences. The presented tool aims to provide natural human-computer multi-modal interaction via exploitation of state-of-the-art technologies in computer vision, speech recognition and synthesis, knowledge representation, sensor data analysis, and by leveraging prior clinical knowledge and patient history through an intelligent, ontology-driven, dialogue manager with reasoning capabilities, which can also access a web search and retrieval engine module. The framework's main contribution lies in its versatility to combine different technologies, while its inherent capability to monitor patient behaviour allows doctors and caregivers to spend less time collecting patient-related information and focus on healthcare. Moreover, by capitalising on voice, sensor and camera data, it may bolster patients' confidence levels and encourage them to naturally interact with the virtual agent, drastically improving their moral during a recuperation process.Entities:
Keywords: Computer vision; Healthcare; Human-computer interaction; Natural language processing; Pervasive systems; Sensors
Year: 2021 PMID: 34127879 PMCID: PMC8190522 DOI: 10.1007/s10844-021-00648-7
Source DB: PubMed Journal: J Intell Inf Syst ISSN: 0925-9902 Impact factor: 1.888
Fig. 1System architecture. We report progress on the components with green background
Fig. 2Bidirectional LSTM-CRF model for Named Entity Recognition
Fig. 3Semantic Web system architecture
Fig. 4Dialogue Management system framework
Doctor-agent use case scenario
| Actor | Interaction |
|---|---|
| (i1) Doctor | REA, did Mr. Goines leave his bed today? |
| (i2) REA | Yes, he walked for 10 minutes, was sitting for 30 minutes and visited the restroom 3 times. |
| (i3) Doctor | Did he have fever during visiting hours? |
| (i4) REA | No, his body temperature was 36.6 degrees Celsius. |
| (i5) Doctor | What is the weather tomorrow? |
| (i6) REA | It will be sunny, around 21 degrees Celsius. |
| (i7) Doctor | Please remind the patient to have a walk at noon. |
| (i8) REA | Reminder set for tomorrow at noon. |
| (i9) Doctor | Thank you, that would be all. |
| (i10) REA | Have a nice day Dr. House! |
Accuracy values for the Samsung S3 mini subset and the LG Nexus 4 subset
| Accelerometer | Gyroscope | DR fusion | Averaging | WACC | ||
|---|---|---|---|---|---|---|
| RF | 0.7225 | 0.6352 | 0.8696 | 0.8709 | 0.8593 | |
| S3 | C5 | 0.7013 | 0.5934 | 0.8131 | 0.8144 | 0.8189 |
| kNN | 0.7887 | 0.5768 | 0.8234 | 0.8491 | 0.8343 | |
| RF | 0.6776 | 0.6426 | 0.7977 | 0.7948 | 0.7948 | |
| Nexus | C5 | 0.6928 | 0.6148 | 0.7727 | 0.774 | 0.7747 |
| kNN | 0.6963 | 0.6131 | 0.7203 | 0.7621 | 0.7582 |
Mean accuracy (%) for different GMM + SFV parameters
Comparison with State-of-the-art
| Method | Accuracy (%) |
|---|---|
| LAFF (SKL) (Hu et al. | 54.2 |
| ST-LSTM (Tree) (Liu et al. | 73.4 |
| Dynamic Skeletons (Hu et al. | 75.5 |
| DPRL + GCNN (Tang et al. | 76.9 |
| Moving Pose + SFV | 74.4 |
Bi-LSTM-CRF settings used for the NER task
| Training parameters | Value |
|---|---|
| Optimiser | Adam |
| Character embeddings dimensions | 100 |
| Word embeddings dimensions | 300 |
| Dropout rate | 0.5 |
| Epochs | 25 |
| Batch size | 20 |
| LSTM size | 100 |
NER performance of the proposed system vs. other popular approaches
| System (CoNLL2003) | Precision | Recall | F1-score |
|---|---|---|---|
| Our system (ELMo embeddings) | 91.63 | 93.01 | 92.32 |
| Our system (BERT embeddings) | 91.46 | 92.32 | 91.88 |
| Lample et al. 2016 | 90.95 | 90.94 | 90.97 |
| Best shared task system: Florian et al. 2003 | 88.99 | 88.54 | 88.76 |
| Baevski et al. 2019 | (not reported) | (not reported) | 93.5 |
Performance evaluation of the proposed system with different embeddings
| System configuration | Correct entities | Wrong entities |
|---|---|---|
| Current embeddings (Glove) | Athens, Smith | Brexit (ORG), Coldplay (LOC), Rea (ORG) |
| With ELMo | Athens, Brexit, Smith | Coldplay (ORG), Rea (ORG) |
| With BERT | Athens, Brexit, Rea, Smith | Coldplay (ORG) |
Sample of the questionnaire used for patient evaluation
| Please rate your agreement on the following statements: | ||||||
|---|---|---|---|---|---|---|
| REA provides the right amount of information. | ||||||
| strongly agree | O | O | O | O | O | strongly disagree |
| REA provides very little information. | ||||||
| strongly agree | O | O | O | O | O | strongly disagree |
| REA provides excessive information. | ||||||
| strongly agree | O | O | O | O | O | strongly disagree |
| REA clearly communicates its intention. | ||||||
| strongly agree | O | O | O | O | O | strongly disagree |
| I always understand what REA is telling me. | ||||||
| strongly agree | O | O | O | O | O | strongly disagree |
| REA understood what I asked for. | ||||||
| strongly agree | O | O | O | O | O | strongly disagree |
| I got the information I wanted. | ||||||
| strongly agree | O | O | O | O | O | strongly disagree |
| REA returns contradictory information. | ||||||
| strongly agree | O | O | O | O | O | strongly disagree |
| REA returns information pertinent to the questions. | ||||||
| strongly agree | O | O | O | O | O | strongly disagree |
| REA works the way I expected. | ||||||
| strongly agree | O | O | O | O | O | strongly disagree |
| REA needs too much time to respond. | ||||||
| strongly agree | O | O | O | O | O | strongly disagree |