| Literature DB >> 32435681 |
Abstract
Gesture recognition technology is rapidly growing in the recent years due to the demands of many application such as computer game and sport, human robot interaction, assistant systems, sign language interpretation and e-commerce. One of the most important of gesture recognition is hand-gesture recognition. For example, it can be used to control all devices (television, radio, air-condition, and doors) by just hand gestures for smart home application. The HGM-4 dataset is built for hand gesture recognition (the full dataset is available from: https://data.mendeley.com/datasets/jzy8zngkbg/4) which contains total 4,160 color images (1280 × 700 pixels) of 26 hand gestures captured by four cameras at different position. The training and testing set are defined to create a benchmark framework for comparing the experimental results.Entities:
Keywords: Biometric recognition; Hand gesture recognition; Image classification; Multiple cameras; One hand gesture; Sign language
Year: 2020 PMID: 32435681 PMCID: PMC7229479 DOI: 10.1016/j.dib.2020.105676
Source DB: PubMed Journal: Data Brief ISSN: 2352-3409
Fig. 1Illustration of one hand-gesture by two different views under different cameras.
Summary of the recent published hand-gesture dataset in the literature.
| Dataset Name | Number of views | Number of gestures | Total images | Resolution | Publicly |
|---|---|---|---|---|---|
| FEMD | 1 | 12 | 1,000 | 640 × 480 | No |
| Interact Play | 2 | 16 | 16,000 | - | Yes |
| IMHG | 2 | 8 | 836 | 640 × 480 | Yes |
| HGM-4 | 4 | 26 | 4,000 | 1280 × 700 | Yes |
The 26 classes of hand gesture of HGM-4 dataset.
| Gesture | Illustration | Gesture | Illustration | Gesture | Illustration |
|---|---|---|---|---|---|
| A | J | S | |||
| B | K | T | |||
| C | L | U | |||
| D | M | V | |||
| E | N | W | |||
| F | O | X | |||
| G | P | Y | |||
| H | Q | Z | |||
| I | R |
Properties of HGM-4 dataset.
| Camera Position | Total images | Number of gestures | Number of images per gesture | Number of acting persons |
|---|---|---|---|---|
| CAM _Left | 1,040 | 26 | 8 | 5 |
| CAM_Right | 1,040 | 26 | 8 | 5 |
| CAM_Front | 1,040 | 26 | 8 | 5 |
| CAM_Below | 1,040 | 26 | 8 | 5 |
Fig. 2Camera setup: each hand-gesture (in front of screen and above the keyboard) is captured at the same time by four cameras.
Fig. 3Illustration of three distinct images of the same gesture captured by below camera.
Fig. 4Original image and its segmentation with removed background by Otsu's method and enhanced by technical expert.
All combination possible to create training and testing set for experiment.
| For one training set | For two training sets | For three training sets | |||
|---|---|---|---|---|---|
| Training | Testing | Training | Testing | Training | Testing |
| CAM _Left | CAM _Right | CAM _Front | CAM _Left | CAM _Right | CAM _Left |
| CAM_Right | CAM _Left | CAM _Left | CAM _Front | CAM _Left | CAM_Right |
| CAM_Front | CAM _Right CAM_Left | CAM_ Below | CAM _Front | CAM _Right | CAM_Front |
| CAM_Below | CAM _Right, CAM_Left, | CAM _Front | CAM_ Below | CAM _Right | CAM_Below |
| CAM _Left | CAM _Right | ||||
| CAM _Right | CAM _Left | ||||
| Subject | Computer Vision, Pattern Recognition, Artificial Intelligence |
| Specific subject area | hand-gesture recognition, image classification, biometric recognition, sign language |
| Type of data | Image (1280 × 700 pixels) in RGB color space |
| How data were acquired | This dataset contains images that were taken by 4 cameras at different positions by laptop camera, indoor condition. |
| Data format | RAW |
| Parameters for data collection | Hand-gesture images are removed background semi-automatically. |
| Description of data collection | This dataset consists of 4,160 images of 26 gestures acquired by 4 different cameras. |
| Data source location | Ho Chi Minh City Open University, Ho Chi Minh City, Vietnam |
| Data accessibility | Mendeley Data |