Literature DB >> 35634047

Efficient 3D AlexNet Architecture for Object Recognition Using Syntactic Patterns from Medical Images.

Shilpa Rani1,2, Deepika Ghai3, Sandeep Kumar4, Mvv Prasad Kantipudi5, Amal H Alharbi6, Mohammad Aman Ullah7.   

Abstract

In computer vision and medical image processing, object recognition is the primary concern today. Humans require only a few milliseconds for object recognition and visual stimulation. This led to the development of a computer-specific pattern recognition method in this study for identifying objects in medical images such as brain tumors. Initially, an adaptive median filter is used to remove the noise from MRI images. Thereafter, the contrast image enhancement technique is used to improve the quality of the image. To evaluate the wireframe model, the cellular logic array processing (CLAP)-based algorithm is then applied to images. The basic patterns of three-dimensional (3D) images are then identified from the input image by scanning the whole image. The frequency of these patterns is also used for object classification. A deep neural network is then utilized for the classification of brain tumor. In the proposed model, the syntactic pattern recognition technique is used to find the feature vector and 3D AlexNet is used for brain tumor classification. To evaluate the performance of the proposed work, three benchmark brain tumor datasets are used, i.e., Figshare, Brain MRI Kaggle, and Medical MRI datasets and BraTS 2019 dataset. The comparative analyses reveal that the proposed brain tumor classification model achieves significantly better performance than the existing models.
Copyright © 2022 Shilpa Rani et al.

Entities:  

Mesh:

Year:  2022        PMID: 35634047      PMCID: PMC9142332          DOI: 10.1155/2022/7882924

Source DB:  PubMed          Journal:  Comput Intell Neurosci


1. Introduction

Object recognition is an open-ended problem in the field of computer vision and medical image processing. For any object to recognize, a human being does not need to do much since we have multiple processing systems with billions of neurons: the brain. We require just a few milliseconds of visual stimulation to recognize an object, but the computer does not work like this because the computer needs more information for object recognition. A whole image can be given as an input but that will have computation limitations. Therefore, a generalized representation of different objects in one domain can give different identities to the same domain's objects. This generalized representation of objects can be further used for the identification of different diseases such as lung cancer and brain tumors. As image processing techniques play an essential role in the diagnosis and monitoring of patients, the biomedical field has gained much attention from researchers, and many advance artificial intelligence-based techniques are available for the early detection of the disease. However, due to the unpredictable nature of cancer, it still needs more advanced techniques. Brain cancer is a life-threatening disease, and approximately every year, 80000 new cases are reported [1, 2]. The treatment of brain cancer always depends on the early and accurate detection of the tumor, and detection always depends on tumor type, type of pathology, and the time of the first investigation of the disease. Generally, brain tumors will irregularly propagate the brain cells [3, 4], and according to the research, three types of brain tumors are identified: glioma, meningioma, and pituitary. Most neurologists will use CAD-based models to detect and classify brain tumors [5]. According to our literature survey, most of the researchers used segmentation-based techniques to classify brain tumors [6-8]. In the last decade, statistical-based classification and detection methods are satisfactory but not a hundred percent accurate [9-13]. Therefore, in this study the image's wireframe model has been developed to detect the brain tumor from the patient's MRI. We proposed an efficient approach based on structural pattern recognition and a deep neural network because structural pattern recognition techniques can represent the patterns of simple and complex images as well. A medical image contains a lot of information, and processing these images is intricate. Hence, a wireframe model is determined and used for further processing. The wireframe model is a skeletal or edge representation of 3D objects using lines and curves [14-17]. Other basic patterns are identified in the 3D image using the syntactic pattern approach, and then, finally, these patterns and frequency of these patterns are calculated for the classification purpose. Deep learning-based algorithms play an essential role in the medical field, and according to the literature survey, the AlexNet-based architecture detects cancer in the early stage more accurately [3, 18–24]. Therefore, the 3D AlexNet architecture is used for the classification of tumor type. This research's primary motivation was to give a new insight into the structural pattern recognition technique because structural pattern recognition techniques can deal with complex medical images [25-30]. Keeping this into mind, we targeted our few objectives, which are as follows: An efficient object recognition model is proposed using syntactic patterns and 3D AlexNet for medical images. Adaptive median filter and contrast image enhancement techniques are used to improve the quality of MRI images. To evaluate the wireframe model of images, the cellular logic array processing (CLAP)-based algorithm is utilized. Basic patterns of three-dimensional (3D) images are also identified by scanning the MRI images. The frequency of these patterns is also used for object classification. The performance of the proposed work is validated on three benchmark brain tumor datasets such as Figshare dataset, Brain MRI Kaggle, and Medical MRI datasets and BraTS 2019 dataset. This study is further divided into four sections to achieve the objectives mentioned above. The second section covers the literature survey. The third and fourth sections cover the proposed methodology and results. The conclusion is discussed in the fifth section.

2. Literature Survey

In the processing of 3D image, superficial and volumetric features play an important role. The edge that intersects each other is referred to as volumetric edges, and the edges that separate two planes are called superficial edges. Lui et al. [6] proposed a gradient-based method for the detection of 3D edges. Rosenfeld et al. proposed an edge detection method using the surface magnitude. The construction of the wireframe model was first proposed by Mukherjee et al. [7]. Thinning and segmentation-based approach is used to evaluate the wireframe model. The adjacent segments are connected using the vertex merging method, and finally, the wireframe model is constructed. Ren et al. [8] implemented a 3D viewpoint and shape estimation-based model to perform wireframe modeling. For the detection of the 3D viewpoints, CNN is used. Prakoonwit and Benjamin et al. [9] proposed 3D surface reconstruction and wireframe modeling by considering 2D images. In this approach, the contours of 2D images are combined and a wireframe model is created. According to the survey, most of the authors have used the viewpoint-based method for the construction of the wireframe model. Much research growth is observed in the field of statistical pattern recognition-based techniques [10], but structural-based recognition still needs researchers' attention. The structural pattern recognition technique can deal with simple and complex patterns, and these intricate patterns can be further divided into sub-patterns. Structural pattern recognition can be used for object recognition, NLP, and scene analysis [11-13]. Ali et al. [13] presented a method for the recognition of Bengali digits. Rashid and Ali [14] proposed a method in which eight directional codes are identified and these direction codes are used to convert the image into a feature vector, but the technique is only applicable for Bengali digits. Pal and Chaudhuri [15] have also proposed a method for the recognition of Bengali numerals. This technique was able to recognize the numerals without thinning and normalization operation. Jaydeb et al. [16] proposed the part of speech (POS) tagging and parsing technique method. This method is used to find the answers from the given text. Hang [17] has presented research on the A algorithm for parsing techniques and proposed an hierarchical graph method (HGM model) based on the same algorithm. This method can replace the existing virtual node method (VNM), which was used to overcome the problem of Chomsky standard forms. The performance of HGM was better than that of VNM. For the last three decades, researchers have extensively used the machine learning approach in various areas, including medical diagnostics. There are limited numbers of studies where researchers specially targeted brain tumor diagnostic problems. Table 1 represents the study of existing methods for brain tumor detection.
Table 1

Study of existing methodology on brain tumor detection.

S. no.Authors and yearMethodologyDatabaseRemarks
1Li et al. [18], 2019Multi-CNNMICCAI BraTS [31]Accuracy = 84.3%
Sensitivity = 82.5%
Specificity = 97.7%

2Addel Gawad et al. [3], 2020Genetic algorithmSelf-datasetFOM = 89.87%
Accuracy = 97.71%
Sensitivity = 94.71%
Specificity = 98.5%

3Mallick et al. [19], 2019Deep wavelet auto-encoder (DWA)Brain image dataset [32]Accuracy = 93.14%
Sensitivity = 94.26%
Specificity = 92.16%
F score = 93.15%

4Shakeel et al. [20], 2019AdaBoost classifier + back propagationSelf-datasetAccuracy = 93.33%
Sensitivity = 71.42%
Specificity = 88.88%

5Anitha and Murugavalli [21], 2015K-means + DWTSelf-datasetAccuracy = 92.43%
Sensitivity = 95.53%
Specificity = 50.6%

6Mano and Anand [22], 2020Swarm intelligence and K-meansFigshare dataset [33]Accuracy = 98.7%
Sensitivity = 93.4%
Specificity = 65.1%
F1 score = 98.7%
JAC = 66.46%
Swati et al. [34], 2019VGG-19Figshare dataset [33]Accuracy = 94.82%

8Khan et al. [31], 2020VGG-16Brain MRI dataset [35]Accuracy = 96%
Precision = 93%
Recall = 100%
F1 score = 97%
Time = 6846sec

9Khan et al. [31], 2020ResNet-50Brain MRI dataset [35]Accuracy = 89%
Precision = 87%
Recall = 93%
F1 score = 90%
Time = 9091sec

10Masood et al. [36], 2021Mask-RCNNBrain MRI dataset [35]Accuracy = 98.34%
11Badza and Barjaktarovic [37], 2020CNNFigshare dataset [33]Accuracy = 95.40%
F1 score = 94.94%
Kurup et al. [38], 2019CapsuleNetFigshare dataset [33]Accuracy = 92.60%
F1 score = 93.33%
Most of the researchers used statistical pattern recognition techniques for object recognition. As per our literature survey, there is no evidence of applying the structural pattern recognition-based method in the medical field. Recently, researchers are starting using deep learning-based algorithms for the segmentation and detection of brain tumors. Hence, we used syntactic patterns for the classification of brain tumors. This is an efficient approach that helps detect and classify brain tumors in the early stage. A detailed description of the proposed methodology is discussed in the next section.

3. Proposed Methodology

This section summarizes the proposed methodology. Inspired by [2], the step-by-step procedure of the proposed classification model is designed (see Figure 1).
Figure 1

Block diagram of the proposed methodology.

3.1. Image Preprocessing

MRI images may contain a different type of noise that can degrade the image's quality and may not provide the required and correct information to detect the cancerous tumor. An efficient denoising and image enhancement technique is required to preserve the edges and the contour of the medical images. Therefore, these techniques are directly helpful in the detection and classification of the image. In the proposed methodology, the adaptive median filter is used for image denoising purposes and an image enhancement technique is used to enhance the image.

3.1.1. Denoising of the Image

Preprocessing techniques are used to remove the noise from the images. Hence, these techniques are directly helpful in the detection and classification of the image. An adaptive median filter is used to remove the noise from the image. It compares each pixel of the image to the neighborhood pixel. If any one of the pixel values is drastically different from the neighborhood pixel, then that pixel will be considered noise and the adaptive median filter will replace that noisy pixel value with the neighborhood pixel's median value. Algorithm 1 of the adaptive median filter is as follows.

3.1.2. Image Enhancement

After removing the noise from the image, the next step is image enhancement. The image enhancement technique is used to improve the image's overall quality and sharpen the edges. In the proposed method, the contrast stretching image enhancement technique is used and is shown in Algorithm 2. This method stretches the range of intensity values, which enhances the image's quality as shown in Figure 2. First, the limit is determined over which intensity value will be extended. These limits vary between 0 and 255 in the 8 bit grayscale image. Further, a histogram of the original image is examined to determine the limits of the original image. If the input image is already covering the full possible set of values, contrast stretching is not required. Otherwise, if input image data are within the restricted range, then starching will be applied. For each image, the original value of r is mapped with the output value O using the following function.
Figure 2

(a) Original image. (b) Denoising image. (c) Contrast-enhanced image.

3.2. Wireframe of the Image

In this step, the preprocessed image is given as an input and then a wireframe model of the image is constructed using the CLAP-based model for 3D images [39-41]. The proposed approach works on the abrupt change in the gray level value of the neighborhood pixels. The algorithm will locate the points over which the gray level is not changing, and it works on all the planes of the 3D image. For the scanning of the 3D image, 3 × 3 × 3 size of the sliding window [42, 43] is used. It is used to verify that some pixels are establishing convex polyhedrons with every sliding window's movement. Thus, at least one pixel in every column and row provides convex polyhedrons. If such type of convex polyhedrons exists, then the central pixel will be updated by zero and the same would be repeated until it does not scan the whole 3D image using a sliding window. The performance of the algorithm is evaluated on various MRI datasets, and it is represented in Figure 3. The step-by-step flow of wireframe model is presented in Algorithm 2.
Figure 3

Wireframe outputs of medical images.

3.3. Knowledge Vector Representation of the Image

There are different types of shape and texture features are available in the image. The knowledge vector representation of the basic 3D patterns [44, 45] gives enough information to classify an object. In this step, we will find the basic pattern of the 3D image, and these 3D patterns will give enough information to classify an object. The following methodology is taken into consideration to identify the pattern in the image syntactic pattern recognition technique.

3.3.1. Finding the Basic Patterns in a 3 × 3 × 3 Space

In a 3D image, a central pixel is surrounded by 26 neighborhood pixels. Hence, the basic patterns have been found by combinations of 26 pixels at a time in a 3 × 3 × 3 space [46, 47]. According to our research, if we consider only the corner pixels of the image then there are 256 possible ways in which a pixel could be present in the 3 × 3 × 3 window. Sample images of the basic 3D patterns are shown in Figure 4. To differentiate these 256 possible combinations/directions, naming conventions are used and the same naming conventions are shown in Figure 4. Here, P1 represents the first pattern, which could be present in the image, and it shows that all the corner pixels are present in the image. Q1 represents that pixel 1 is not present in the 3 × 3 × 3 size of the window. In the same way, R1,3 represents that corner pixels 1 and 3 are not present in the 3 × 3 × 3 size of the window. R1,7 represents that corner pixels 3 and 7 are not present in the 3 × 3 × 3 size of the window. We need to find these patterns in the image and then calculate the frequency of the image.
Figure 4

Sample images of basic patterns present in the 3D image.

The naming convention is defined for all 256 patterns. Figure 5 shows the naming convention of the particular pixel position. Figures 5(a) and 5(b) represent 27-neighborhood structure of 3 × 3 × 3 window and co-ordinates, respectively. If we consider a 3 × 3 × 3 size of an image, then a center pixel (14) can be surrounded by 26 pixels, but in our research, we considered only corner pixels, which are supposed (1, 3, 7, 9, 19, 21, 25, and 27). Hence, the possible way in which a corner pixel could be present is as follows:
Figure 5

(a) 27-neighborhood structure of 3 × 3 × 3 window. (b) Representation of co-ordinates.

The possible way in which a corner pixel could be present in the image is shown in Table 2.
Table 2

Representation of a possible combination of convex polyhedrons.

GroupsNumber of pixels eliminated
Group PNil. Possible combination—1P1 = {1,3,7,9,19,21,25,27}
Group QOne of the corner pixels is eliminated. Possible combination—8Q1 = {3,7,9,19,21,25,27},Q3 = {1,7,9,19,21,25,27}, Q7 = {1,3,9,19,21,25,27},Q9 = {1,3,7,19,21,25,27}, Q19 = {1,3,7,9,21,25,27},Q21 = {1,3,7,9,19,25,27}, Q25 = {1,3,7,9,19,21,27}, Q27 = {1,3,7,9,19,21,25}
Group RAny two corner pixels are eliminated. Possible combination—28R1,3 = {7,9,19,21,25,27}, R1,7 = {3,9,19,21,25,27}
R1,9 = {3,7,19,21,25,27}, R1,19 = {3,7,9,21,25,27}
R1,21 = {3,7,9,19,25,27}, R1,25 = {3,7,9,19,21,27}
R1,27 = {3,7,9,19,21,25}, R3,7 = {1,9,19,21,25,27}
R3,9 = {1,7,19,21,25,27}, R3,19 = {1,7,9,21,25,27}
R3,21 = {1,7,9,19,25,27}, R3,25 = {1,7,9,19,21,27}
R3,27 = {1,7,9,19,21,25}, R7,9 = {1,3,19,21,25,27}
R7,19 = {1,3,9,21,25,27}, R7,21 = {1,3,9,19,25,27}
R7,25 = {1,3,9,19,21,27}, R7,27 = {1,3,9,19,21,25}
R9,19 = {1,3,7,21,25,27}, R9,21 = {1,3,7,19,25,27}
R9,25 = {1,3,7,19,21,27}, R9,27 = {1,3,7,19,21,25}
R19,21 = {1,3,7,9,25,27}, R19,25 = {1,3,7,9,21,27}
R19,27 = {1,3,7,9,21,25}, R21,25 = {1,3,7,9,19,27}
R21,27 = {1,3,7,9,19,25}, R25,27 = {1,3,7,9,19,21}
Group SAny three pixels are eliminated. Possible combination—56Same way for these groups also
Group TAny four pixels are eliminated. Possible combination—70
Group UAny five pixels are eliminated. Possible combination—56
Group VAny six pixels are eliminated. Possible combination—28
Group WAny seven pixels are eliminated. Possible combination—8
Group XAll corner pixels are eliminated. Possible combination—1 X 1,3,7,9,19,21,25,27 = { }

3.3.2. Run a 3 × 3 × 3 Window on an Input Image and Check for the Patterns

The 2D slices can be combined, and it can form a 3D structure/3D volumetric data. The proposed feature extraction algorithm is written in such a way that it can work with 2D slices and 3D volumetric datasets. The proposed method takes input as a 3D volume or a sequence of 2D frames (e.g., slices in a CT scan). If it is a 2D slice and the scanning window size is 3 × 3, then a central pixel could be surrounded by eight neighborhood pixels. The naming conventions of the preferred direction are as follows: Right(R),DownRight(DR),Down(D),DownLeft(DL),Left(L),Upleft(UL),Up(U), and UpRight(UR) If we consider the sequence of 2D slices that can form a 3D structure and again the sliding window size is 3 × 3 × 3, then the central pixel could be surrounded by 26 neighborhood pixels. The naming conventions of the preferred direction are as follows: Here, F represents front and B represents backward direction. Other conventions are the same as 2D. The wireframe model of the image is considered as an input for this step, and the detailed explanation is given as follows. (1) Finding the Patterns in the Image and Calculating their Frequency Using a Stride of3. Consider a whole image and a 3 × 3 × 3 window. This window slides over the image and finds out the pattern in which this 3 × 3 × 3 window encapsulates. If any portion of the image consists of the patterns, then the frequency value of that particular pattern would be updated. To go to the next 3×3×3 section of the image, the window moves with a stride of three. This procedure is repeated until the end of the image has been reached. The three-step stride may necessitate padding around the edges of an image in order to retrieve its information. It is to be noted that the stride of 1 is in X, Y, and Z directions. When one 3 × 3 × 3 row of an image is slid on, the next 3 × 3 × 3 row is obtained by sliding the window 3 pixels in the downward direction. (2) Finding the Frequency of All the Patterns Obtained (3D Pattern Frequency Vector). The patterns calculated in the previous step will be counted using the vocabulary of 256 patterns, and their frequency is calculated. The frequencies of all the patterns will be represented as an array. This array will be called 3D pattern frequency vector or 3D-PFV in short. The length of this vector will be 256, which corresponds to the frequency of the patterns. The frequency of the patterns written in the array should be the same and consistent. If pattern 2 is on index 1 of the 3D-PFV for one image, then it should be on the same index for all the images. The same goes for all the patterns' frequencies. This 3D-PFV representation of the image will be a mapping between the vector and the object. 3D AlexNet input layer will be a column stacked 3D picture feature vector (3D-PFV). The hidden layer will have n neurons with n weights, and the output layer will have m neurons for  m classes of objects.

3.3.3. Object Recognition Using 3D Pattern Frequency Vector

The three-dimensional knowledge vector is a vector of 256 values that saves an image's overall syntactic pattern. This image can be two- or three-dimensional. The purpose of such a vector is to get a succinct representation of a 2D or 3D image. These vectors, therefore, hold essential information regarding the shape and morphology. Therefore, theoretically, these vectors hold information about an image and thus can be used to represent it. The corresponding vector is generated using the research shown in the previous section. For each syntactic pattern, the image is calculated and fed as an integer to the 3D AlexNet. Since there are 256 patterns, the 256-dimensional vector is created, tentatively called the X-dimensional pattern frequency vector (XDPFV), where X is 3. In the proposed methodology, 3D AlexNet is used for classification purposes. AlexNet is implemented in 2012 by Alex Krizhevsky. 3D AlexNet architecture consists of 3D filters at the convolution layer and pooling layer instead of 2D filters. In the 3D AlexNet, eight layers are present, of which 5 are convolution layers and 3 are fully connected layers. 1st convolution layer is followed by the first max-pooling layer and the same for the second convolution layer. 3rd layer is connected to consecutive positions, and then, the 5th convolution layer is connected to the 3D max-pooling layer. The output generated by the 3rd max-pooling layer will become the input to the next two fully connected layers. The third fully connected layer will now be input for the soft-max classifier, and the soft-max layer consists of three class labels. The architecture of 3D AlexNet is represented in Figure 6.
Figure 6

3D AlexNet architecture.

4. Results and Discussion

In this research, we used four different brain MRI datasets. The first dataset is “Figshare brain tumor dataset,” which is publicly available [33]. This dataset consists of a total of 3064 images of 233 different patients. The second dataset is again “Brain MRI dataset,” which is publicly available on Kaggle [35]. This dataset consists of a total of 253 MRI images. The third dataset is the Medical MRI dataset, and it is collected from the Pentagram Research Institute, Hyderabad. The dataset contains 2D slices of MRI, which can be combined to make a 3D view of the images. The fourth dataset is BraTS 2019 dataset [48]. This dataset consists of the images of four different contrasts. The dataset contains four folders. The folder structure is composed of fluid-attenuated inversion recovery (FLAIR), T1, T1 contrast-enhanced, and T2. To build a multi-orientation or view of each 3D model, we rotated the dataset equally in a horizontal direction and finally produced the dataset on 4th and 7th depths and multi-orientation (3, 6, 9, 12, 18, 24, and 30 rotations) occurs only on depth 7th. Each 3D model's multi-orientation is applied to both the training and testing datasets individually. To train the deep learning model, Keras with TensorFlow libraries has been used. The specification of the system is Windows Operating System, GPU Processor, and 32 GB RAM. We created a GUI of the proposed system for better analysis purposes, which takes MRI images as input. GUI of the proposed methodology is shown in Figure 7.
Figure 7

GUI of the proposed work.

During the brain tumor classification, 3D AlexNet architecture is used, which requires values of different hyper-parameters to be set. This is required to obtain the optimum performance of the architecture. These hyper-parameters are epochs, learning rate, dropout, and batch size. The optimized values of these hyper-parameters for each problem are mentioned in Table 3. Learning rate and early stopping depend on the validation loss for every 5 to 10 epochs. In this implementation, the batch size is considered 128. The epoch is 65, and steps per each epoch are 4.
Table 3

Details of 3DAlexNet classifier.

Parameter nameValues
Epochs65
Learning rate0.001
Dropout0.20
Batch size128
No. of convolution8
No. of fully connected layer3
Pooling layerMax pooling
Pooling layer window size2
The early stopping property has been introduced in the training process, which stops the learning process at certain epochs; in case, the accuracy and loss are constant for continuous four epochs. The validation split is set to 0.1. This parameter allows the model to handle train/test data splitting on its own. 0.1  splits the data by splitting the sample dataset into 90% of training and 10% to test the samples. This allows plotting the loss and accuracy values correctly while plotting them. The dropout technique used to ignore a random set of neurons during the training is given as 0.20. The dropout set to 20% means one in 5 inputs is excluded from each cycle. The learning rate ranges from 0 to 1. The learning rate that is too high can cause the model to converge too swiftly to a suboptimal solution. If the learning rate is too low, it causes the model to get stuck between the training processes. Simultaneously, a learning rate that is too small can cause the process to get stuck. In the proposed method, we have taken the learning rate as 0.001 so that the network learns slowly to perform object recognition on divergent input images.

4.1. Evaluation Parameters

The proposed methodology's performance is evaluated by the computation of accuracy, sensitivity, specificity, and F1 score defined in Table 4 [47, 49]. If accuracy, sensitivity, and specificity have a higher value, it represents the algorithm's better performance.
Table 4

List of evaluation parameters and their formula.

S. No.Parameters nameFormula
1AccuracyAccuracy=TP+TN/(TP+TN+FP+FN)
2SpecificitySpecificity=TN/TN+TP
3SensitivitySensitivity=TP/TP+FN
4F1 scoreF1 − score=2precisionrecall/precision+recall

TP = true positive, TN = true negative, FP =false positive, and FN =false negative.

4.2. Result Analysis

MRI images were selected to evaluate the performance of the proposed system. First, a comparative analysis of the proposed wireframe model with various classical edge detection algorithms has been done, and for the comparison purpose, the Sobel, Robert, and Prewitt algorithms are taken into consideration. According to the result analysis, it is observed that the Robert and Sobel edge detection method could not produce the closed contour, leading to false tumor detection. The proposed wireframe model can generate a closed structure of the tumor and boundary of the tumor is sharp and visibility is more.

4.2.1. Qualitative Analysis

Table 5 shows the subjective quality evaluation among various methods. As shown in Table 5, the proposed wireframe model is able to find the contour/wireframe model for the small object also, which is not possible in the case of Robert, Sobel, and Prewitt. As per the analysis, the proposed CLAP-based wireframe model performs better than existing methods.
Table 5

Comparative analysis of various edge detection methods on the different 3D views of the estimated tumor inside a cancerous brain.

Original imageWireframeRobertSobelPrewitt

4.2.2. Quantitative Analysis

The results obtained from the proposed wireframe model are compared with the existing edge detection method on the test images shown in Table 6. The performance measures are accuracy, sensitivity, and specificity. The accuracy of the proposed wireframe model is 99%, the sensitivity is 87%, and the specificity of 1 shows the proposed wireframe model's effectiveness for detecting the edges from the given images. The detailed statistics measures are shown in Figures 8–10.
Table 6

Performance analysis of various edge detection methods with the proposed wireframe method.

MethodAccuracySensitivitySpecificity
Proposed0.990.871
Robert0.820.150.85
Sobel0.810.140.85
Prewitt0.850.190.90
Figure 8

Comparative analysis of various edge detection methods with the proposed wireframe method.

Figure 9

Comparative analysis of network parameters of existing systems and proposed system on Figshare dataset.

Figure 10

Comparative analysis of the proposed method with RCNN-based approach on the Figshare dataset.

4.3. Result of Classification on Different Datasets

This section represents the comparative analysis of the classification results of the proposed method with the existing methods over the same datasets. For the classification of different types of tumors, different hyper-parameters were used for the training of the proposed model. The obtained results are evaluated using different quantitative measures, which are already mentioned in Table 6. The performance of the proposed method is evaluated on the Figshare dataset [33], Brain MRI Kaggle dataset [35], and Medical MRI dataset and BraTS 2019 dataset. The evaluation parameters are accuracy, sensitivity, specificity, precision, recall, F1 score, and mean average precision (mAP). Table 7 shows the comparative analysis of the classification results obtained by existing methods on the Figshare dataset [33]. The highest accuracy is obtained using the fine-tune VGG-16 [61] model. The authors have used fine-tuning transfer learning, which is used to improve the efficiency of the architecture. Figure 9 shows the graphical representation of the proposed and existing method results on the Figshare dataset.
Table 7

Result analysis of the proposed work with other state-of-art methodology on Figshare dataset.

MethodModelAccuracy (%)
Cheng et al. [50]SVM87.54
Dheng et al. [51]SVM94.68
Abir et al. [52]PNN83.33
WIdhiarso et al. [53]CNN82
Afshar and Plataniotis [54]CapsNet86.56
Abiwinanda et al. [55]CNN84.19
Ismael and Qader [56]MPNN91.90
Pashaei and Jazayeri [57]ELM93.68
Deepak and Ameer [58]GoogleNet and SVM97.10
Swati et al. [34]VGG-1994.82
Huang et al. [59]CNN based on complex networks95.49
Gumaei et al. [60]GIST descriptor and ELM94.93
Chakrabarty [35]Attention module, hyper column technique, residual block97.69
Arisha Rehman et al. [61]Fine-tune AlexNet97.39
Arisha Rehman et al. [61]Fine-tune GoogleNet98.04
Arisha Rehman et al. [61]Fine-tune VGG-1698.69
Proposed methodology3D AlexNet 99.04
Some authors have used the RCNN-based method for the classification of the brain tumor, and Table 8 shows the comparative analysis of the existing method on the Figshare dataset. The evaluation parameters are accuracy, mAP, sensitivity, and time. As per the comparative study, the proposed methods give better results in terms of accuracy and time on the Figshare dataset. Figure 10 shows the graphical representation of the proposed and existing method results on the Figshare dataset.
Table 8

Performance comparison of the proposed method with existing RCNN-based methods on Figshare dataset.

MethodModelEvaluation metrics
AccuracymAPSensitivityTime (s)
Rayahan et al. [62]RCNN0.9200.9100.9500.47
Ren et al. [63]Faster RCNN0.9400.9400.9400.25
Masood et al. [36]ResNet-500.9590.9460.9530.20
Masood et al. [36]DenseNet-410.9630.9490.9530.20
Proposed method 0.990 0.9580.9850.13
For a better evaluation of the proposed method, we performed the experiments on two other datasets. The accuracy of the proposed method on the brain MRI dataset is 99.17%. As per the comparative study, mask RCNN performs better with an accuracy of 98.34% , but our proposed system has 1.14% higher accuracy rate than the existing system. Some other evaluation parameters are used for the validation of the proposed work. The result of the comparative analysis is shown in Table 9. The class-wise accuracy of the Figshare dataset is 98.94%, 99.18%, and 99.02% on glioma, meningioma, and pituitary classes. The accuracy of the proposed method is also evaluated on the Medical MRI dataset, and it is 99.23%. The results of the comparative analysis are shown in Table 10.
Table 9

Result analysis of the proposed work with other state-of-art methodology on the Figshare dataset.

MethodModelClassesAccuracy (%)Sensitivity (%)Specificity (%)F1 score (%)
Rehman et al. [61]Freeze AlexNet91.2190.4895.48
Rehman et al. [61]Freeze GoogleNet93.4695.3897.74
Rehman et al. [61]Freeze VGG-1689.7687.8194.64
Pashaei et al. [64]CNN93.6893.00
Badza and Barjaktarovic [37]CNN95.4094.94
Kurup et al. [38]CapsuleNet92.6093.33
Srinivasan et al. [65]GLCM and wavelet packets93.3072.00
Proposed methodGlioma98.9498.9498.7997.3
Meningioma99.1898.1199.0198.7
Pituitary99.0298.7298.1998.69
Table 10

Result analysis of the proposed work with other state-of-art methodology on the Brain MRI Kaggle dataset and Medical MRI dataset.

MethodModelAccuracyPrecisionRecallF1 scoreTime (sec)
Swati et al. [34]VGG-190.94
Masood et al. [36]Mask-RCNN0.98
Hassan et al. [31]CNN1.01.01.01.03085
Hassan et al. [31]VGG-160.960.931.00.976846
Hassan et al. [31]ResNet-500.890.870.930.909091
Hassan et al. [31]Inception-v30.750.770.710.745630
Proposed work on Brain MRI dataset 3D AlexNet 0.99 0.970.940.913706
Proposed work on Medical MRI dataset 3D AlexNet 0.99 0.980.950.933891
The performance of the proposed work was evaluated on the BraTS 2019 dataset. Table 11 shows the average accuracy of the proposed method on the BraTS 2019 dataset. This section compares the proposed method with seven other existing methods, and all the researchers worked on two-class labels: LGG and HGG tumors. Zhugeet et al. [70] used deep CNN for the classification of tumor using T1, T2, and T2-FLAIR images and achieved the impressive accuracy of 97.1%. The proposed method used 3D AlexNet for the classification of tumor using T1, T1ce, T2, and T2-FLAIR images and achieved the accuracy of 96.91%, which is 0.19% lesser than the existing method [70]. Table 12 shows the class-wise performance of the various method using precision, specificity, recall, and F1 score. For low-grade glioma, the proposed 3D AlexNet achieved the highest precision of 0.925. This value is 0.062% higher than the existing pretrained ResNet mixed convolution method. For high-grade glioma and healthy subject class, the 3D AlexNet method achieved the highest precision of 0.959 and 0.998, which are higher than the existing ResNet mixed convolution method. In the same way, the performance of the proposed method in terms of recall for LGG and HGG classes is better than the existing one, but in the case of healthy subjects the recall value of the proposed method is 0.956, which is 0.039 less than the pretrained ResNet mixed convolution method. As with LGG, HGG, and healthy subjects, the best model is the exiting ResNet mixed convolution with the specificity of 0.931, 0.959, and 0.999. The F1 score value of the proposed method for LGG and HGG classes is better than the existing method, but in the classification of healthy subjects pretrained ResNet 3D achieved the highest F1 score value.
Table 11

Result analysis of the proposed work on the BraTS 2019 dataset.

StudyMethodContrastAccuracy (%)
Shahzadi et al. [66]CNN with LSTMT2-FLAIR84.00
Pei et al. [67]CNNT1, T1ce, T2, T2-FLAIR74.9
Ge et al. [68]Deep CNNT1, T2, T2-FLAIR90.87
Mzoughi et al. [69]Deep CNNT1, T1ce, T2, T2-FLAIR96.59
Zhuge et al. [70]Deep CNNT1, T2, T2-FLAIR97.1
Ouerghi et al. [71]Random forestT1, T2, T2-FLAIR96.5
Chatterjee et al. [72]Pretrained ResNet mixed convolutionT1ce96.98
Proposed3D AlexNetT1, T1ce, T2, T2-FLAIR96.91
Table 12

Result analysis of the proposed work with other state-of-art methodology on the BraTS 2019 dataset.

StudyMethodClassesPrecision (%)Recall (%)Specificity (%)F1 score (%)
Chatterjee et al. [72]ResNet 3DLow-grade glioma0.7980.9200.9070.854
High-grade glioma0.9330.8280.9620.877
Healthy subjects0.9930.990.990.99

Chatterjee et al. [72]Pretrained ResNet 3DLow-grade glioma0.7810.8550.9110.814
High-grade glioma0.8960.8350.9370.863
Healthy subjects1.000.9991.000.999

Chatterjee et al. [72]ResNet(2 + 1)DLow-grade glioma0.7860.9140.9020.844
High-grade glioma0.9300.8220.9620.873
Healthy subjects0.9940.9900.9970.992

Chatterjee et al. [72]Pretrained ResNet(2 + 1)DLow-grade glioma0.8410.9100.9310.873
High-grade glioma0.9280.8700.9590.897
Healthy subjects0.9990.9980.9990.999

Chatterjee et al. [72]ResNet mixed convolutionLow-grade glioma0.7470.8860.8550.773
High-grade glioma0.9110.7500.9630.823
Healthy subjects0.9940.9760.9970.985

Chatterjee et al. [72]Pretrained ResNet mixed convolutionLow-grade glioma0.8630.9310.9120.894
High-grade glioma0.9440.8830.9360.912
Healthy subjects0.9970.9950.9960.996

Proposed 3D AlexNet Low-grade glioma0.9250.9460.9630.904
High-grade glioma0.9590.8950.9980.919
Healthy subjects0.9980.9560.9780.968

5. Conclusion and Future Work

In this research, a CLAP-based algorithm was used for the detection of the wireframe model. The algorithm took various input images, and their corresponding wireframe model was computed. The wireframe model finds the close surface of the given tumor image and also provides the 3D visual representation of the images. The qualitative and quantitative comparisons were performed on various edge detection algorithms. The accuracy of the CLAP-based wireframe model was found to be 99%, which is better than the majority of the existing models. Further, tumor detection and classification-based system were developed using a wireframe model and syntactic pattern recognition approach. Experiment analysis reveals that the proposed lightweight brain tumor classification model has achieved comparatively better performance than the existing state-of-art methods. Furthermore, the proposed system could combine with the existing cancer diagnostic tools to improve the robustness of the existing CAD-based systems.
  13 in total

1.  An analysis of some common scanning techniques for lossless image coding.

Authors:  N Memon; D L Neuhoff; S Shende
Journal:  IEEE Trans Image Process       Date:  2000       Impact factor: 10.856

2.  Brain tumor classification for MR images using transfer learning and fine-tuning.

Authors:  Zar Nawab Khan Swati; Qinghua Zhao; Muhammad Kabir; Farman Ali; Zakir Ali; Saeed Ahmed; Jianfeng Lu
Journal:  Comput Med Imaging Graph       Date:  2019-05-18       Impact factor: 4.790

3.  Brain tumor classification using deep CNN features via transfer learning.

Authors:  S Deepak; P M Ameer
Journal:  Comput Biol Med       Date:  2019-06-29       Impact factor: 4.589

4.  Multi-modality medical image fusion technique using multi-objective differential evolution based deep neural networks.

Authors:  Manjit Kaur; Dilbag Singh
Journal:  J Ambient Intell Humaniz Comput       Date:  2020-08-08

5.  Brain tumor classification in MRI image using convolutional neural network.

Authors:  Hassan Ali Khan; Wu Jue; Muhammad Mushtaq; Muhammad Umer Mushtaq
Journal:  Math Biosci Eng       Date:  2020-09-15       Impact factor: 2.080

6.  Deep Learning and Multi-Sensor Fusion for Glioma Classification Using Multistream 2D Convolutional Networks.

Authors:  Chenjie Ge; Irene Yu-Hua Gu; Asgeir Store Jakola; Jie Yang
Journal:  Annu Int Conf IEEE Eng Med Biol Soc       Date:  2018-07

7.  Enhanced Performance of Brain Tumor Classification via Tumor Region Augmentation and Partition.

Authors:  Jun Cheng; Wei Huang; Shuangliang Cao; Ru Yang; Wei Yang; Zhaoqiang Yun; Zhijian Wang; Qianjin Feng
Journal:  PLoS One       Date:  2015-10-08       Impact factor: 3.240

8.  A Novel Deep Learning Method for Recognition and Classification of Brain Tumors from MRI Images.

Authors:  Momina Masood; Tahira Nazir; Marriam Nawaz; Awais Mehmood; Junaid Rashid; Hyuk-Yoon Kwon; Toqeer Mahmood; Amir Hussain
Journal:  Diagnostics (Basel)       Date:  2021-04-21

9.  Classification of brain tumours in MR images using deep spatiospatial models.

Authors:  Soumick Chatterjee; Faraz Ahmed Nizamani; Andreas Nürnberger; Oliver Speck
Journal:  Sci Rep       Date:  2022-01-27       Impact factor: 4.379

Review 10.  The Multimodal Brain Tumor Image Segmentation Benchmark (BRATS).

Authors:  Bjoern H Menze; Andras Jakab; Stefan Bauer; Jayashree Kalpathy-Cramer; Keyvan Farahani; Justin Kirby; Yuliya Burren; Nicole Porz; Johannes Slotboom; Roland Wiest; Levente Lanczi; Elizabeth Gerstner; Marc-André Weber; Tal Arbel; Brian B Avants; Nicholas Ayache; Patricia Buendia; D Louis Collins; Nicolas Cordier; Jason J Corso; Antonio Criminisi; Tilak Das; Hervé Delingette; Çağatay Demiralp; Christopher R Durst; Michel Dojat; Senan Doyle; Joana Festa; Florence Forbes; Ezequiel Geremia; Ben Glocker; Polina Golland; Xiaotao Guo; Andac Hamamci; Khan M Iftekharuddin; Raj Jena; Nigel M John; Ender Konukoglu; Danial Lashkari; José Antonió Mariz; Raphael Meier; Sérgio Pereira; Doina Precup; Stephen J Price; Tammy Riklin Raviv; Syed M S Reza; Michael Ryan; Duygu Sarikaya; Lawrence Schwartz; Hoo-Chang Shin; Jamie Shotton; Carlos A Silva; Nuno Sousa; Nagesh K Subbanna; Gabor Szekely; Thomas J Taylor; Owen M Thomas; Nicholas J Tustison; Gozde Unal; Flor Vasseur; Max Wintermark; Dong Hye Ye; Liang Zhao; Binsheng Zhao; Darko Zikic; Marcel Prastawa; Mauricio Reyes; Koen Van Leemput
Journal:  IEEE Trans Med Imaging       Date:  2014-12-04       Impact factor: 10.048

View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.