Literature DB >> 32372771

An efficient image descriptor for image classification and CBIR.

Ashkan Shakarami1, Hadis Tarrah2.   

Abstract

Pattern recognition and feature extraction of images always have been important subjects in improving the performance of image classification and Content-Based Image Retrieval (CBIR). Recently, Machine Learning and Deep Learning algorithms are utilized widely in order to achieve these targets. In this research, an efficient method for image description is proposed which is developed by Machine Learning and Deep Learning algorithms. This method is created using combination of an improved AlexNet Convolutional Neural Network (CNN), Histogram of Oriented Gradients (HOG) and Local Binary Pattern (LBP) descriptors. Furthermore, the Principle Component Analysis (PCA) algorithm has been used for dimension reduction. The experimental results demonstrate the superiority of the offered method compared to existing methods by improving the accuracy, mean Average Precision (mAP) and decreasing the complex computation. The experiments have been run on Corel-1000, OT and FP datasets.
© 2020 Elsevier GmbH. All rights reserved.

Entities:  

Keywords:  Content-based image retrieval (CBIR); Convolutional neural network (CNN); Image classification; Image descriptor; Machine learning; Pattern recognition

Year:  2020        PMID: 32372771      PMCID: PMC7198219          DOI: 10.1016/j.ijleo.2020.164833

Source DB:  PubMed          Journal:  Optik (Stuttg)        ISSN: 0030-4026            Impact factor:   2.443


Introduction

In recent years, because of the wide spread use of the Internet and the massive use of audio-visual information in digital format for communications, designing the systems for describing the content of multimedia information in order to seek and classify them is really important. In computer vision, image descriptors describe elementary characteristics such as shape, color, texture or motion of images, which are visual features of images [1,2]. Offering new image descriptors has been active research area and will help in increasing the performance of many tasks in computer vision. Some descriptors such as HOG, LBP, SURF and SIFT have been proposed, so far [[3], [4], [5], [6], [7]]. These descriptors are usually used in Machine Learning for pattern recognition and feature extraction [[8], [9], [10], [11]]. Each of these descriptors has disadvantages, like large dimension of feature vector and considering only certain features such as texture. To overcome these problems, an efficient image descriptor is presented in this research. This descriptor is created using combination of HOG, LBP and improved AlexNet CNN. Furthermore, for dimension reduction the PCA has been applied. When the proposed descriptor is used for image classification and CBIR, it provides benefits which are divided into following items: Higher accuracy and mAP compare with other works; Overcoming the problem of high dimension descriptors such as HOG; Sensitivity to intra-class as well as inter-class variety; High performance on imbalanced databases. The rest of this paper is organized as: The related works is introduced in section 2. The proposed method is discussed in section 3. The experimental results are reported in section 4 and these results are compared with existing experimental results in this section. Finally, the conclusion and future work are presented in section 5.

Related works

In recent years, image descriptors are usually used for many cases such as image classification and CBIR [12,13]. Some researchers utilized combination of descriptors to propose new ones in order to increase efficiency and use it for special cases [14]. Some of them are described as follow: [15] has offered a descriptor using combination of HOG and LBP. The results of this approach have showed that, it is an efficient descriptor for object detection, but the unequal dimension is the main problem of that. This descriptor is affected by HOG. Suppose the dimension of input image is 227 × 227 × 3; in this case, the feature vector’s dimension of HOG descriptor is 1 × 26244, whereas this dimension for LBP descriptor is 1 × 59 and the combined descriptor has more affected by the larger descriptor. Whatever this difference become larger, the smaller descriptor lose its effect [16,17]. [18] similar to [15], has suggested a way that feature vector’s dimension of HOG and LBP descriptors (1 × 181,535) are combined and are reduced to 1 × 2824 using the variance and improved PSO algorithm, in which LBP descriptor has more influence. In our proposed method, for overcoming mentioned problems a solution is presented for dimension reduction using PCA algorithm. Convolutional Neural Networks can be used for feature extraction, and the extracted features are utilized in cases such as classification, retrieval or detection. [19] proposed a method in which the fusion of extracted features by two networks used for face detection. In some works the combination of Convolutional Neural Networks and descriptors is used to create a new image descriptor. For instance [20], has suggested a method that image features are extracted using the HOG and SURF descriptors, then the features of each descriptor are sent to convolutional layers, and the feature vectors with 1 × 2016 and 1 × 1024 dimensions are created. After that, these features are combined and their dimension is reduced by Fully Connected (FC) layers and eventually are used for classification. In this way, such as [15,18], because of the inequality of the dimensions of HOG and SURF descriptors, their impacts on the outcome of classification will be different. Considering the weaknesses of the mentioned existing methods, in this study a new descriptor has been offered to overcome these problems.

Proposed method

In Fig. 1 , the schematic of proposed method has been shown. According to this figure, in the first step, the images are read and resized to 227 × 227 × 3. Then, each image is sent to deep feature extractor (an improved AlexNet CNN) and handcrafted descriptors such as HOG and LBP simultaneously. On the one hand, the improved AlexNet CNN processes the images and recognizes its patterns, and finally proposes a feature vector with the dimension of 1 × 64 [21]. On the other hand, the HOG and LBP descriptors extracts features. Then the PCA algorithm is used to reduce the dimensions of produced features by HOG descriptor.
Fig. 1

Proposed method.

Proposed method. In this paper, for having equal effects of HOG and LBP descriptors on final feature descriptor, the dimensions of HOG descriptor has been reduced to 1 × 59 by PCA algorithm. After that, the HOG-PCA and LBP feature’s vectors are combined and a new handcrafted feature vector with dimension of 1 × 118 is created. In order to match the dimension of handcrafted feature vector with deep feature vector the PCA is applied on handcrafted feature vector and 64 of 118 features are selected and a Handcrafted-PCA feature vector with a dimension of 1 × 64 is created. Finally, the deep feature vector and Handcrafted-PCA feature vector are combined and an efficient image descriptor with a dimension of 1 × 128 is created. In this research, the AlexNet CNN, HOG, LBP and PCA are chosen by authors due to some reasons. These reasons are described as follow: The Feature selection technique which is applied by PCA algorithm, has been used for several reasons such as reducing the computation volume and training times, simplification of models and etc. [22]. The LBP descriptor is a strong feature for texture classification. Combining the LBP with HOG descriptor, improves the performance of detection considerably on some datasets [15,23]. The HOG descriptor is computed on a dense grid of uniformly spaced cells and for improving the accuracy, uses overlapping local contrast normalization [62]. The advantages of this descriptor is that it executes on local cells which is invariant to geometric and photometric transformation, except for object orientation [4,24]. The improved AlexNet CNN detects the important and high level features automatically without any human supervision [21].

Simulation and comparison results

The proposed method has been implemented with Matlab 2018b software, a computer system with 6GB RAM, a NVIDIA GeForce 920 M graphics processor unit (GPU), and an Intel® Core™ i5-7200U @ 2.50 GHz central processor unit (CPU). In this investigation, the accuracy, mean average precision (mAP) and recall criteria have been measured in order to evaluate the proposed method for classification and CBIR [[25], [26], [27], [28]] and 3-fold cross-validation has been applied in all experiments (Westerhuis et al 2008). In addition, the Adam algorithm has been used to train the AlexNet CNN [29] and Random forest, SVM and KNN classifiers have been utilized for classification [[30], [31], [32], [33], [34], [35], [36]]. Moreover, for measuring the similarity, Euclidean distance has also been used [37,38]. In the following, used datasets, experiments and comparison results have been demonstrated and described in details.

Used datasets

In this study, for evaluating the offered way, Corel-1000 (Wang), OT and FP datasets have been used. In Table 1 , the number of categories and number of images per categories have been shown. These datasets are described in details as below:
Table 1

The number of categories and images per categories for used datasets.

DatasetsNumber of categoriesNumber of images
Corel-1000101000
OT82688
FP (Catlech-101)5380
The number of categories and images per categories for used datasets.

Corel-1000 dataset

Corel-1000 dataset contain 10 different categories including Africa, Beaches, Building, Bus, Dinosaur, Elephant, Flower, Horses, Mountain and Food. In this dataset, each category include 100 images which are with the size of 256 × 384 or 384 × 256 pixels [39].

Oliva and Torralba (OT) dataset

OT dataset include 8 categories and 2688 images: 260 highway, 292 streets, 360 coasts, 328 forest, 308 inside of cities, 374 mountain, 410 open country and 356 tall buildings, in which, forest is considered for all forest and rivers scenes. Because almost all of the images include the sky object, there is not a specific sky scene. Most of the scenes present a large inter-class variability, however this annotations make a higher inter-class variability besides a large intra-class variability. In addition, this dataset is an imbalanced dataset [40].

FP (Caltech-101) dataset

FP dataset consist of 5 categories of Caltech-101 dataset include Bonsai, Joshua tree, Lotus, Sunflower and Water Lilly which contain 128, 64, 66, 85 and 37 images, respectively. Therefore, this dataset like OT is an imbalanced dataset [[41], [42], [43]].

Simulations and evaluation of proposed method for image classification

In this section, the performance of proposed method using Corel-1000, OT and FP databases has been evaluated and compared to AlexNet CNN for image classification [61]. For better comparison all below experiments have been done in same condition.

Simulation and evaluation of proposed method on Corel-1000 dataset for image classification

In Table 2 , the results of AlexNet CNN on Corel-1000 dataset for image classification have been demonstrated. According to this table, the accuracy of AlexNet CNN in train data is 97.80% and the accuracy in test data is 90.10%. Also, standard divisions of 3-fold for train and test data are 0.46 and 1.80, respectively. Moreover, by considering Table 3 , it is understood that, this method offered higher accuracy in training and test phase compared with AlexNet CNN. In presented method, Random forest, SVM and KNN have been evaluated for classification. As can be seen, Random forest classifier has made higher accuracy on test data which is 6% more than AlexNet CNN.
Table 2

Evaluation of AlexNet CNN on Corel-1000 dataset for Image Classification.

MethodTrain data (Accuracy ± Standard division)Test data (Accuracy ± Standard division)
Accuracy ± Standard divisionAccuracy ± Standard division
AlexNet CNN97.80 ± 0.4690.10 ± 1.80
Table 3

Evaluation of proposed method on Corel-1000 dataset for Image Classification.

MethodTrain data (Accuracy ± Standard division)
Test data (Accuracy ± Standard division)
Classifiers
Classifiers
Random forestSVMKNNRandom forestSVMKNN
Proposed method10010010096 ± 0.6294.70 ± 1.0895.20 ± 1.04
Evaluation of AlexNet CNN on Corel-1000 dataset for Image Classification. Evaluation of proposed method on Corel-1000 dataset for Image Classification.

Simulation and evaluation of proposed method on OT dataset for image classification

In Table 4 , the results of AlexNet CNN on OT dataset for image classification have been illustrated. As can be seen, the accuracy of AlexNet CNN on test data is 89.17%. Also, standard divisions of 3-fold for test data is 0.48. In Table 5 , the results of proposed method shows that the accuracy is 93.86% on test data that is 4% more than AlexNet CNN.
Table 4

Evaluation AlexNet CNN on OT dataset for Image Classification.

MethodTrain data (Accuracy ± Standard division)Test data (Accuracy ± Standard division)
Accuracy ± Standard divisionAccuracy ± Standard division
AlexNet CNN95.67 ± 0.6189.17 ± 0.48
Table 5

Evaluation proposed method on OT dataset for Image Classification.

MethodTrain data (Accuracy ± Standard division)
Test data (Accuracy ± Standard division)
Classifiers
Classifiers
Random forestSVMKNNRandom forestSVMKNN
Proposed method10099.67 ± 0.0310093.86 ± 0.1793.86 ± 0.2193.19 ± 0.09
Evaluation AlexNet CNN on OT dataset for Image Classification. Evaluation proposed method on OT dataset for Image Classification.

Simulation and evaluation of proposed method on FP dataset for image classification

In Table 6 , the results of AlexNet CNN on FP for image classification have shown that the accuracy of test data in AlexNet CNN is 88.95%. In Table 7 , the results of the proposed method have been shown. The results show that the accuracy of proposed method on test data is 89.74% and could achieve higher accuracy compare with AlexNet CNN. In all these experiments, Random forest classifier is more accurate than other classifiers.
Table 6

Evaluation AlexNet CNN on FP dataset for Image Classification.

MethodTrain data (Accuracy ± Standard division)Test data (Accuracy ± Standard division)
Accuracy ± Standard divisionAccuracy ± Standard division
AlexNet CNN98.16 ± 0.8888.95 ± 1.55
Table 7

Evaluation proposed method on FP dataset for Image Classification.

MethodTrain data (Accuracy ± Standard division)
Test data (Accuracy ± Standard division)
Classifiers
Classifiers
Random forestSVMKNNRandom forestSVMKNN
Proposed method10099.87 ± 0.1010089.74 ± 0.8888.16 ± 0.8289.74 ± 0.52
Evaluation AlexNet CNN on FP dataset for Image Classification. Evaluation proposed method on FP dataset for Image Classification.

Simulations and evaluation of proposed method for Contend-based Image Retrieval (CBIR)

In this section, the performance of proposed method using Corel-1000, OT and FP databases has been evaluated and compared to AlexNet CNN for CBIR. All below experiments have been done in same condition.

Simulation and evaluation of proposed method on Corel-1000 dataset for CBIR

In Table 8 , the results of AlexNet CNN and proposed method on Corel-1000 dataset for CBIR have been demonstrated. According to this table, the proposed method has a higher mAP for 5-top, 10-top and all relative images retrieval than AlexNet CNN. Consequently, this method could achieve better performance compare with a base model such as AlexNet CNN. Also the dimension of proposed feature vector is 1 × 128, whereas the AlexNet CNN has a feature vector with 1 × 4096 dimension for CBIR [44].
Table 8

Evaluation proposed method and comparison with AlexNet CNN on Corel-1000 dataset for CBIR.

MethodsmAP ± Standard division
5-top10-topAll relative images
AlexNet CNN93.1491.8775.48
Proposed method96.02 ± 1.9495.80 ± 1.8291.65 ± 1.22
Evaluation proposed method and comparison with AlexNet CNN on Corel-1000 dataset for CBIR.

Simulation and evaluation of proposed method on OT dataset for CBIR

According to Table 9 and by comparing the proposed method with AlexNet CNN on OT dataset for CBIR, it is found that the proposed method is more efficient and has higher mAP.
Table 9

Evaluation proposed method on OT dataset for CBIR.

MethodsmAP ± Standard division
5-top10-topAll relative images
AlexNet CNN93 ± 0.4892.30 ± 0.6571.26 ± 1.58
Proposed method94.22 ± 0.3693.91 ± 0.3785.64 ± 0.55
Evaluation proposed method on OT dataset for CBIR.

Simulation and evaluation of proposed method on FP dataset for CBIR

Table 10 illustrates, the presented way has a higher mAP for 5-top, 10-top and all relative images retrieval compare with AlexNet CNN on FP dataset that leads to better performance.
Table 10

Evaluation proposed method on FP dataset for CBIR.

MethodsmAP ± Standard division
5-top10-topAll relative images
AlexNet CNN83.78 ± 2.3481.80 ± 2.2271.23 ± 2.36
Proposed method87.58 ± 1.4486.86 ± 1.2083 ± 0.50
Evaluation proposed method on FP dataset for CBIR.

The proposed method’s mAP-Mean recall plots

In this section, the mAP-mean Recall plots of mentioned method on Corel-1000, OT and FP databases for CBIR have been illustrated. In Plot 1 , the efficiency of proposed method are shown on Corel-1000 dataset. According to this plot, the mAP and mean recall for 5-top retrieval are 96.02% and 4.5%, respectively. These values for 10-top retrieval are 95.80% and 8.25%. Furthermore, for all relative images are 91.65% and 50.75%. By considering these results, it can be said that:
Plot 1

The mAP-Mean recall plot of the proposed method for Corel-1000 dataset.

The offered methodology can retrieve many relevant images perfectly. The mAP-Mean recall plot of the proposed method for Corel-1000 dataset. The efficiency of proposed method on OT dataset are depicted in Plot 2 . According to this plot, the mAP and mean Recall for 5-top retrieval are 94.22% and 1.36%, for 10-top retrieval are 93.91% and 2.5% and for all relative images are 85.64% and 50.23% respectively. Hence, for this dataset, the presented technique could retrieve many relevant images too.
Plot 2

The mAP-Mean recall plot of the proposed method for OT dataset.

The mAP-Mean recall plot of the proposed method for OT dataset. The efficiency of proposed method for FP dataset, are demonstrated in the Plot 3 . According to this plot, the mAP and mean Recall for 5-top retrieval are 87.58% and 6.97%, respectively. These values for 10-top retrieval are shown 86.86% and 12.77% and for all relative images are 83% and 51.16%. Therefore this a successful method in retrieving relevant images.
Plot 3

The mAP-Mean recall plot of the proposed method for FP dataset.

The mAP-Mean recall plot of the proposed method for FP dataset.

Visual representation of proposed method's retrieval performance

In this section, visual representation of proposed technique in term of CBIR is illustrated. For each dataset and its categories, an image is randomly selected and content based image retrieval is done. Finally, the 5-top retrieve images are shown.

Visual representation of proposed method's retrieval for Corel-1000 dataset

The visual results of proposed method for CBIR on Corel-1000 dataset are depicted in Table 11 . For each category, one query image is randomly picked out. After similarity measurement, 5 more similar images to query image are retrieved and illustrated.
Table 11

The visual results of proposed method for CBIR on Corel-1000 dataset.

The visual results of proposed method for CBIR on Corel-1000 dataset. According to Table 11, it can be concluded that, in more cases the most similar images to query image have been retrieved and placed in first position that is what is expected from a good CBIR system. For example, in Africa category, an African native presents in query image and in retrieved images African natives with high similarity to query image are presented. Also, in Beaches category, the retrieved images have beach, sea, sky and people similar to query image. In addition, in other category such as Dinosaur, the retrieved dinosaurs have the most similarity to query image and more of them belong to same species. It means: The proposed method is sensitive to inter-class variety as well as intra-class variety and can find and retrieve the most similar images, despite the diversity within the class [45,46]; Considering the horse and flower category, it is found that the presented method can take into account features such as color and shape besides texture features [[47], [48], [49]].

Visual representation of the proposed method's retrieval for OT dataset

In Table 12 , the visual results of proposed method for CBIR on OT dataset are demonstrated. According to this table, it can be found that:
Table 12

The visual results of proposed method for CBIR on OT dataset.

Although this dataset is imbalanced and the imbalanced datasets may lead to problems, the proposed method is able to retrieve similar images in imbalanced dataset too [50,51]. The visual results of proposed method for CBIR on OT dataset.

Visual representation of proposed method's retrieval for FP dataset

In Table 13 , the visual results of proposed method for CBIR on FP dataset are shown. According to this table, it can be figured out that although this dataset has categories in which images have similar contents (Such as Lotus and Water Lilly), it performs perfectly in retrieving images [43,52].
Table 13

The visual results of proposed method for CBIR on FP dataset.

The visual results of proposed method for CBIR on FP dataset.

Comparison results

In Table 14 , the results of suggested way in term of classification compared with other available methods is demonstrated. Considering this table, it is concluded that the proposed method has higher accuracy compared with other existing methods. Therefore the new image descriptor is an effective descriptor for image classification.
Table 14

Comparison results of proposed method with other method for image classification.

DatasetMethodAccuracyTraining rate%, Test rate%
Corel-1000Fuzzy Topological [53]62.2033, 67
Color Histogram + Fuzzy Neural Network [54]73.40
Proposed method963-fold cross validation
OTFusion features [55]63.89
Co-occurance matrix + Bayesian classifier [43]8910-fold cross validation
Hybrid generative + Dense SIFT [56]91.0850, 50
Proposed method93.863-fold cross validation
FPCo-occurance matrix + KNN [43]87.2010-fold cross validation
Proposed method89.743-fold cross validation
Comparison results of proposed method with other method for image classification. In Table 15 , the retrieval results of mentioned method has been shown. According to this table, this method has good performance in CBIR, too. Moreover, it is more efficient because of higher mean average precision (mAP), whereas the dimension of proposed method is same to references [[65], [66], [67]].
Table 15

Comparison results of proposed method with other method for CBIR.

DatasetMethodmAPTraining rate%, Test rate%Dimension of proposed method
Corel-1000CCM + DBPSP [63]76.1090, 101 × 21
Block Truncation Coding [57]77.901 × 96
HOG + SURF [64]80.6170, 30
Dense SIFT [65]84.2050, 501 × 128
SURF + FREAK [66]8670, 301 × 128
SURF + MSER [67]8870, 301 × 128
Fusion features [55]83.50
AlexNet CNN [44]93.801 × 4096
Proposed method95.803-fold cross validation1 × 128
OTCo-occurance matrix [43]76.3910-fold cross validation1 × 9
Color moment + Angular Radial Transform + Edge histogram [58]50.5985, 15
Relevance Feedback [59]79
Proposed method93.913-fold cross validation1 × 128
FPCo-occurance matrix [43]78.8310-fold cross validation1 × 9
Proposed method86.863-fold cross validation1 × 128
Comparison results of proposed method with other method for CBIR. The feature vector of each image in offered method is 1 × 128. Suppose a single byte is required to work on any element of 1 × 128. So, in this case only 128 bytes of memory for each vector is required. However the dimension of feature vector become larger, more memory is needed. For example, 4096 bytes of memory were needed to measure the similarity of each image when was used FC7 in the paper of [44]. Therefore, the suggested method is more effective than these studies [18,60].

Conclusion and future works

In this research, an efficient method for image description has been proposed. According to the results of experiments and comparing the proposed method to existing works, it is obvious that the proposed method is an efficient way for image description hence, it is suitable for image classification and CBIR. One of the merits of proposed method is sensitivity to intra-class and inter-class variety when it is used for CBIR. In other words, related images with more similarity are retrieved early. Another advantage is high performance on imbalanced databases. Plus, the presented way could reduce high dimension of descriptors such as HOG using PCA and has increased the accuracy and mAP. Some of current researches have focused on designing and development of Computer-Aided diagnosis (CAD) systems. In these systems, data descriptors play vital role. Because of the epidemic of COVID-19 coronavirus and its dangerous, if a descriptor can extract important and key features on medical images of patients (such as CT), it will be used on designing an efficient CAD system for early diagnosis of this disease. Thus, extending the proposed method on CAD system can be a target for future researches. Furthermore, using new and powerful convolutional neural networks such as EfficientNet and MobileNetV3 instead of AlexNet CNN can be valuable in future studies. The authors of this research are also completing their studies in this regard.

Declaration of Competing Interest

No conflict of interest.
  8 in total

1.  On the Euclidean distance of images.

Authors:  Liwei Wang; Yan Zhang; Jufu Feng
Journal:  IEEE Trans Pattern Anal Mach Intell       Date:  2005-08       Impact factor: 6.226

2.  Face description with local binary patterns: application to face recognition.

Authors:  Timo Ahonen; Abdenour Hadid; Matti Pietikäinen
Journal:  IEEE Trans Pattern Anal Mach Intell       Date:  2006-12       Impact factor: 6.226

3.  Scene classification using a hybrid generative/discriminative approach.

Authors:  Anna Bosch; Andrew Zisserman; Xavier Muñoz
Journal:  IEEE Trans Pattern Anal Mach Intell       Date:  2008-04       Impact factor: 6.226

4.  A completed modeling of local binary pattern operator for texture classification.

Authors:  Zhenhua Guo; Lei Zhang; David Zhang
Journal:  IEEE Trans Image Process       Date:  2010-03-08       Impact factor: 10.856

5.  Content based medical image retrieval based on new efficient local neighborhood wavelet feature descriptor.

Authors:  Amita Shinde; Amol Rahulkar; Chetankumar Patil
Journal:  Biomed Eng Lett       Date:  2019-05-06

Review 6.  Measuring the accuracy of diagnostic systems.

Authors:  J A Swets
Journal:  Science       Date:  1988-06-03       Impact factor: 47.728

7.  Driver Drowsiness Detection Based on Steering Wheel Data Applying Adaptive Neuro-Fuzzy Feature Selection.

Authors:  Sadegh Arefnezhad; Sajjad Samiee; Arno Eichberger; Ali Nahvi
Journal:  Sensors (Basel)       Date:  2019-02-22       Impact factor: 3.576

8.  An effective content-based image retrieval technique for image visuals representation based on the bag-of-visual-words model.

Authors:  Safia Jabeen; Zahid Mehmood; Toqeer Mahmood; Tanzila Saba; Amjad Rehman; Muhammad Tariq Mahmood
Journal:  PLoS One       Date:  2018-04-25       Impact factor: 3.240

  8 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.