| Literature DB >> 30417045 |
Marco Colasito1, Jeremy Straub2, Pratap Kotala2.
Abstract
This data set is comprised of correlated audio and lip movement data in multiple videos of multiple subjects reading the same text. It was collected to facilitate the development and validation of algorithms used to train and test a compound biometric system that consists of lip-motion and voice recognition. The data set is a collection of videos of volunteers reciting a fixed script that is intended to be used to train software to recognize voice and lip-motion patterns. A second video is included of the individual reciting a shorter phrase, which is designed to be used to test the recognition functionality of the system. The recordings were collected in a controlled, indoor setting with a 4K professional-grade camcorder and adjustable, LED lights.Entities:
Year: 2018 PMID: 30417045 PMCID: PMC6218630 DOI: 10.1016/j.dib.2018.10.043
Source DB: PubMed Journal: Data Brief ISSN: 2352-3409
Fig. 1Diagram of subject, camera and lighting position.
Fig. 2A subject is recorded while reading from the script.
Fig. 3Depicts three frames from the video, showing how they can be processed to detect the subject’s face and lip positions and movement.
| Subject area | |
| More specific subject area | |
| Type of data | |
| How data was acquired | |
| Data format | |
| Experimental factors | |
| Experimental features | |
| Data source location | |
| Data accessibility | |
| Related research article |