| Literature DB >> 28796781 |
Hiroki Tanaka1, Hideki Negoro2, Hidemi Iwasaka3, Satoshi Nakamura1.
Abstract
Social skills training, performed by human trainers, is a well-established method for obtaining appropriate skills in social interaction. Previous work automated the process of social skills training by developing a dialogue system that teaches social communication skills through interaction with a computer avatar. Even though previous work that simulated social skills training only considered acoustic and linguistic information, human social skills trainers take into account visual and other non-verbal features. In this paper, we create and evaluate a social skills training system that closes this gap by considering the audiovisual features of the smiling ratio and the head pose (yaw and pitch). In addition, the previous system was only tested with graduate students; in this paper, we applied our system to children or young adults with autism spectrum disorders. For our experimental evaluation, we recruited 18 members from the general population and 10 people with autism spectrum disorders and gave them our proposed multimodal system to use. An experienced human social skills trainer rated the social skills of the users. We evaluated the system's effectiveness by comparing pre- and post-training scores and identified significant improvement in their social skills using our proposed multimodal system. Computer-based social skills training is useful for people who experience social difficulties. Such a system can be used by teachers, therapists, and social skills trainers for rehabilitation and the supplemental use of human-based training anywhere and anytime.Entities:
Mesh:
Year: 2017 PMID: 28796781 PMCID: PMC5552034 DOI: 10.1371/journal.pone.0182151
Source DB: PubMed Journal: PLoS One ISSN: 1932-6203 Impact factor: 3.240
Fig 1System framework of role-playing and feedback through interaction with computer avatar.
Fig 2Extracted facial landmark points using face tracker.
Fig 3Audiovisual feedback provided by automated social skills trainer: User video, overall score, comparison with models, and positive comments.
Fig 4Improvement of overall narrative scores in audio and audiovisual groups.
Error bars indicate standard error (*: p<.05).
Fig 5Overall narrative scores of pre- and post-training.
Participants are indicated by color. We added a small amount of noise to separate identical points.
Relationship between participant ages and scores.
| ID | Age | (a) Pre | (b) Post | (b) Post—(a) Pre |
|---|---|---|---|---|
| 1 | 7 | 1 | 4 | 3 |
| 2 | 7 | 3 | 5 | 2 |
| 3 | 9 | 1 | 5 | 4 |
| 4 | 11 | 5 | 6 | 1 |
| 5 | 12 | 4 | 6 | 2 |
| 6 | 12 | 6 | 6 | 0 |
| 7 | 12 | 6 | 6 | 0 |
| 8 | 13 | 4 | 6 | 2 |
| 9 | 16 | 5 | 6 | 1 |
| 10 | 19 | 4 | 5 | 1 |
Correlation coefficient to overall narrative skills.
| Non-verbal skill categories | Correlation coefficient |
|---|---|
| Structure | 0.93 |
| Speech rate | 0.92 |
| Gesture | 0.90 |
| Smoothness | 0.89 |
| Posture | 0.85 |
| Fillers | 0.84 |
| Intonation | 0.81 |
| Smile | 0.80 |
| Difficult words | 0.80 |
| Amplitude | 0.79 |
| Eye contact | 0.78 |
| Face orientation | 0.73 |