ISSN: 0970-938X (Print) | 0976-1683 (Electronic)

Biomedical Research

An International Journal of Medical Sciences

Research Article - Biomedical Research (2016) Volume 27, Issue 4

Unified RF-SVM model based digital radiography classification for Inferior Alveolar Nerve Injury (IANI) identification

Manikandaprabhu P* and Karthikeyan T

PSG College of Arts and Science, India

*Corresponding Author:
Manikandaprabhu P
Department of Computer Science
PSG College of Arts & Science
Coimbatore
Tamilnadu, India

Accepted date: March 30, 2016

Visit for more related articles at Biomedical Research

Abstract

This paper proposes a feature selection based classification method that can be applied to better accuracy in the Digital Radiographs (DR) for the identification of Inferior Alveolar Nerve Injury (IANI). Different conventional features based on the shape and EZW (Embedded Zero tree Wavelet) based texture features are extracted using different feature extraction techniques. Then, aggregate voting of all the decision trees in the Random Forest (RF) method to get the reduced feature. Finally, Support vector machine (SVM) with RBF kernel is trained using reduced features to classify the IAN (Inferior Alveolar Nerve) injured on a healthy object. The proposed classification results are compared with PCA-DT, PCA-MLP and PSO-MLP classification algorithms. Both training and testing stages of proposed model get better classification accuracy of 96.4% and 83.58% respectively. This shows the highest classification accuracy performance among some other existing methods.

Keywords

Digital radiographs, Feature selection, Inferior alveolar nerve injury, Random forest, Support vector machine.

Introduction

Medical demands have driven the fast growth of medical image processing methods, which in turn have significantly promoted the level of the therapeutic process. They have become essential tools for diagnosis, treatment planning and authentication of administered treatment. Therefore, the technology of medical image processing has long attracted attention of relevant experts. Nowadays, panoramic radiography is regularly used in dental practice, because it provides visualness of anatomical structures in pathological changes of the teeth, jaws and temporo-mandibular joints.

A dental implant is surgically inserted into the jawbone as an artificial root onto which a dental prosthesis is placed. Inferior Alveolar Nerve injury (IANI) is the main problematical significance of dental surgical procedures and has major medico-legal implications [1]. These injuries results in fractional or absolute loss of consciousness from the ipsilateral skin of the lower lip and cheek, the buccal oral mucosa in this area and the lower teeth. Causes of IANI include placement of dental implants, local anesthetic injections, third molar surgery, endodontic, trauma and orthognathic surgery.

IANI and Lingual Nerve (LN) injuries are considered to be suitable to mechanical irritations from surgical interference and are inclined by numerous demographic, anatomic and treatment-related factors [2]. Most patients who extend an IANI gradually get back to regular consciousness above the path of a few weeks or months, depending upon the nerve damage severity. Nevertheless, after the most rigorous injuries, where fraction/fracture or all of the nerves have been sectioned and the place may have been compromised by infection, revival will be imperfect. Therefore, identification of the IAN is essential when using preoperative planning software for dental implant surgery.

The IAC has a hollow tubular shape but is difficult to detect because its structure is not well defined and because it is frequently connected to nearby hollow spaces. Hence, accurate and automatic identification of the IAN canal is a challenging issue. The IAC is outlined by two thin radiopaque outlines to signify the cortical walls of the canal. IAC appears lower or superimposed above the apex of the mandibular molar teeth [4]. Thus, the majority of earlier researches on the IAC detection [5] require users’ intervention such as a manual trace of the IAN canal. Damage to the IAN can yield very serious complications [3].

On panoramic Digital Radiography (DR) observation, the root apex of the mandibular second molar was in close propinquity to the mandibular canal while the apices of the mesial and distal roots of the mandibular first molar were the farthest from the canal [6]. Radiological prediction of injury to the IAN depends on the relationship of the root to the canal discussed by Malik [7]. Rood and Shebab [8] has defined seven vital recommendations that can be taken from OPG images. There have been several OPG assessment researches which maintain the effectiveness of these seven findings [9]. It must be renowned, even if, that the statistical results from these analysis had various levels of specificity and sensitivity in Table 1. Nowadays, research decided that high-risk signs were known by OPG in particular; darkening of the root is closely related to cortical bone loss and/or grooving of the root [9].

OPG Imaging procedure Sensitivity (%) Specificity (%)
Darkening of the root 32-71 73-96
Interruption of the canal 22-80 47-97
Diversion of the canal 3-50 82-100

Table 1. Specificity and sensitivity for predicting IAN injury with third molar surgery in OPG image [9].

biomedres-Anatomy-IAN

Figure 1. Anatomy of IAN.

Image classification [10] automatically assigns an unknown image to a category according to its visual content, which has been a major research direction in computer vision. Feature extraction step affects all other subsequent processes. Visual features were categorized into primitive features such as colour, shape and texture. However, as X-ray images are gray level images and do not contain any colour information, the related image analysis systems mostly deal with textures for feature extraction process which were used by several researchers [1114]. A multi-visual feature mixture such as GLCM, shape based canny edge detector feature and pixel value was presented by Mueen et al. [15]. Other authors have also combined pixel value as a global image descriptor with other image representation techniques to construct feature vector of the image [16].

Texture analysis has a major role for medical image analysis. Gray level co-occurrence matrices (GLCM) [17] are commonly used feature extraction techniques in the above works. GLCM histogram has extracted the feature vector used for Digital Radiograph (DR) analysis by Yu.et al. [11]. Texture based geometric pattern feature analysis method to analyze the DR developed by Katsuragawa et al. [12]. GLCM and embedded zero tree wavelet features are extracted using PCADT classification method in [13]. PCA-MLP classification method has proposed in [14] using embedded zero tree wavelet and GLCM features. Here, Random Forest [RF] is used as feature selection method [18]. From a classification point of view, support vector machine (SVM) is a popular method for binary classification. Zhu et al. [19] proposed SVM with Gaussian kernel classifier evaluate the benign and malignant pulmonary modules. Yuan et al. [20] assess the breast cancer prognosis classification using SVM with better accuracy than other classifiers. It automatically identifies the medical related messages using SVM model outperformed [21]. Therefore, in this work SVM is used as a classification tool.

Proposed Methodology

Proposed Methodology framework shown in Figure 2 is as follows:

biomedres-Proposed-Methodology

Figure 2. Proposed Methodology

• Preprocessing

• Feature Extraction

• Feature Selection

• Classification

Preprocessing

Preprocessing consists of normalizing the intensity variations, low contrast, removing low frequency noise, removing reflections and masking segment of images. This step is a major role in the perfect and precise extraction of features. Histogram adjustment, noise removal and canny based edge extraction are sublevel of this process.

Histogram adjustment: It is used to emphasize and highlight the details of input images and to justify their gray level histogram [22]. In this case, the contrast of all images is increased by mapping the values of the input intensity image of new values such that, 1% of the data is saturated at low and high intensities of the input data (Figure 3).

biomedres-OPG-image

Figure 3. OPG image before preprocessing.

Noise removal: To get rid of noise and unnecessary information, the anisotropic diffusion filter as proposed in [23] is utilized. This filter removes the noise from input image while preserves important parts of image contents such as edges and major boundaries. The filtered image is modeled as the solution to the anisotropic diffusion equation as follows:

equation (1)

where equation is a scale image and is a decreasing function depending on the gradient of u .

Edge and boundary extraction: An edge gives us an object layout. Identifying the pixel value and it is compared with the neighboring pixels to be outlined as edge regions [24]. All objects in the image are outlined when the intensities are measured exactly [25]. In order to extract the edges and boundaries from images, canny edge detection algorithms [24] is utilized. In Canny edge detection procedure for segmentation shows in Figure 4.

biomedres-Edge-Boundary-Detection

Figure 4. OPG image for After Edge and Boundary Detection.

Feature extraction

Mostly, image classification use low level features color, shape and texture. But, DR images are mainly in grey level. So, color is not suitable feature for DR analysis. Then, shape and texture can be considered for DR analysis [12]. Here, we extracted the conventional features in Figure 5 as follows: Shape, Region properties, Tamura features, entropy and Wavelet based Texture Features for two-Dimensional level discrete wavelet transform for four (LL, LH, HL and HH) sub-bands in θ directions(θ=0°, 45°, 90°, 135°). Here, there are 383 features are extracted from each DR image.

biomedres-Conventional-features

Figure 5. Conventional features.

Shape features: Shape provides geometrical information of an object in an image. This geometrical information is remained same even when the location, scale and orientation of the object are changed. In this study, the shape information of an image is described based on its edges. A histogram of the edge directions is used to represent shape attribute for each image and image patches.

Edge direction histogram (EDH): For each pixelequation its edge vector is defined as image whereequation and equation are obtained by the vertical edge mask and horizontal edge mask respectively. The edge vector can be also represented as its amplitude and direction, the amplitude can be roughly estimated by

equation (2)

The angle representing the pixel’s edge direction is decided by

equation (3)

Each pixel in the image has an edge vector containing its edge amplitude and direction, and all the pixels’ edge vectors form the image’s edge map. It must be noted that in the actual implementation of the algorithm, equation (3) is not necessary. In this paper, the equation is used to express the pixel’s edge direction.

The edge direction histogram [26] is used to compare the edge information between the distorted image blocks and the reference ones. In this paper, the continual direction (0°~180°) is divided into 8 discrete directions which are shown in Figure 6 and are not the same as reference [26].

biomedres-discrete-directions

Figure 6. The 8 discrete directions.

For each image block, its edge direction histogram can be obtained by the following steps:

Step 1. Calculate each pixel’s edge amplitude and direction by equations (2) and (3) respectively.

Step 2. Quantify each pixel’s direction as one of the 8 discrete directions (see Figure 6).

Step 3. Sum up all the pixels’ edge amplitudes with the same direction in the block.

Now edge direction histogram (vector) is obtained. An example of a 16 × 16 image block and its edge direction histogram are showed in Figure 7.

biomedres-edge-direction-histogram

Figure 7. A 16 × 16 image block and its edge direction histogram, D represents the directions and AMP represents the amplitude.

Wavelet-texture features: Various statistical features are derived from the segmented image for classification [17]. But it is a challenging task to extract a good feature set for classification. GLCM features are based on the joint probability distribution of matchup of pixels. Distanced and angle ө within a given neighborhood are used for computation of joint probability distribution between pixels. Normally d=1, 2 and ө=0°, 45°, 90°, 135° are used for computation [27]. Hence, 32 features totally obtained each DR image. In our proposed system the following GLCM features are extracted [28]:

Energy (k1) which expresses the repetition of pixel pairs of an image,

equation (4)

Local variations (k2) which present in the image is calculated by Contrast. If the contract value is high, the image has large variations.

equation (5)

Correlation (k3) is a measure linear dependency of grey level values in co-occurrence matrices. It is a two dimensional frequency histogram in which individual pixel pairs are assigned to each other on the basis of a specific, predefined displacement vector.

equation (6)

Homogeneity (k4) is inversely proportional to contrast at constant energy whereas it is inversely proportional to energy.

equation (7)

Wavelet transform can capture both frequency and spatial information and has merits of multi-resolution and multi-scale decomposition [29,30]. The selection of the wavelet base is very important in practical applications of DWT. When analyzing the same problem, a different wavelet basis will produce a different result. So far, there has been no good way or a unified standard to solve the problem, and the main method used is to determine a wavelet base that is applicable for a problem according to the error between the results derived from wavelet analysis and the theoretical results.

In our study, we selected the well-known Zero tree Wavelet as the wavelet base, which may provide a more effective analysis than others in the scenario of image processing. In order to perform the DWT, the image has to be a square image, and its row/column size must be an integer power of 2. So, technically, the EZW is applicable to the square images of sizes in integer powers of 2 (for example, image sizes like 128 x 128 or 512 x 512). This transformation is theoretically lossless, although this may not always be the case. The purpose of the transformation is to generate decorrelated coefficients, which means it removes all the dependencies between samples.

The EZW algorithm applies Successive Approximation Quantization in order to provide multi-precision representation of the transformed coefficients and to facilitate the embedded coding. The algorithm codes the transformed coefficients in decreasing order in several scans. Each scan of the algorithm consists of two passes: significant map encoding and refinement pass. The dominant pass scans the subband structure in zigzag, right-to-left and then top-to-bottom within each scale, before proceeding to the next higher scale of subband structure as presented in Figure 8.

biomedres-EZW-subband-structure

Figure 8. EZW subband structure scanning order.

After decomposition for the first time, the original image is divided into four sub-band images of equal size of that quarter. These four images are LLi, LHi, HLi and HHi, which represent the ith scale. The sub-band LLi represents the low-frequency sub-image, corresponding to an approximation image, which is to be decomposed in the (i+1)th scale; the sub-bands LHi, HLi and HHi collectively called low-frequency sub-images correspond to detail images. In first-level decomposition, the size of the original image of size M×N is decomposed into four sub-bands LL1, LH1, HL1 and HH1, of size M/2×N/2. After the ith-level decomposition, the size of the three sub-bands LHi, HLi and HHi is M/2i×N/2i [31].

Region properties: The common region properties [32] are perimeter, eccentricity, Euler number, area, major axis length, minor axis length and orientation of the image. The minor and major axis lengths are lengths of the minor and major axes of the ellipse that has the same normalized second central moments as the region or an object respectively. The eccentricity can be defined as the ratio of the smallest eigenvalue to the largest one. The Euler number is the subtraction of number of connected components and number of holes in the object or region. The orientation can be defined as the direction of the largest eignvector of the second order covariance matrix of a region or an object. The perimeter is the number of pixels located on the object boundary and the area is the number of pixels located in the region within the boundary.

Tamura features: There are 6 different Tamura features: coarseness, contrast, directionality, linelikeness, regularity and roughness [33]. In the literature [34], the first three features are used since they are strongly correlated with human perception.

Entropy: It is a general statistical parameter. Log-Energy entropy, Sure entropy and Shanon entropy [35], shown in Equation (8), are generally used to extract feature [36].

equation

equation

equation (8)

Where equationand equation is the image with size equation is a constant where equation N is the total number of pixels in image I and “#” denotes the number of elements in the set. As the information theoretical features like Entropy, depend on gray level intensities, the images gray level should be normalized before feature extraction until exact and accurate features are generated. Hence, to normalize the gray level intensities, Equation (9) is utilized. If equation is the original image pixel, the normalization equation is:

equation (9)

Where equation and equation are the initial (desired) mean and variance of original image I, respectively and equation is the normalized image. In this paper, the desired mean and variance are selected 100 and 2000 respectively. After the normalization, three different types of Entropy features are extracted which provide a three element feature vector.

Feature selection

Feature selection [18] retains higher accuracy at the use of minimal feature subset. It can be accomplished through wrapper and filter methods. Wrappers depend heavily on classification algorithm to measure the prominence of a feature to be included in the model. Feature selection through wrappers generally performs better than filters because the filter selection is optimized for the particular learning algorithm to be used. Wrapper methods are very time taking and they are computationally expensive. Here, Random Forest (RF) algorithm is used as a feature selection method [18].

Random forest (RF): It is the most popular machine learning method, but it also provides feature importance for their relatively good accuracy, robustness and ease of use [37]. It consists of a number of decision trees. Building many decision tree predictors with randomly selected variable subsets and utilizing a different subset of training and validation data for each individual model. After generating many trees, aggregate voting of all trees in the forest based class prediction occurred. Thus, lower ranked variables are eliminated based on empirical performance heuristics [18]. The main advantages of RF are: handles numerous inputs attributes - both qualitative and quantitative, estimate relatively important features for feature selection, fast learning and modest computation time. In Table 2, shown the Number of features selected based on the feature selection methods.

Feature Selection Methods Number of Features Selected
Before Feature Selection 383
PCA 41
PSO 45
RF 35

Table 2. Summary of Feature Selection.

Classification

In mid-1990s, Vapnik [38] introduced Support Vector Machine (SVM). It performs well in small sample sizes and in high dimensional spaces [19]. It also exploits the least feasible quadratic programming issues, which determines systematically and promptly. It improves the scaling and computing time significantly. It executes the lowest sequential maximization procedure to train a classifier using linear kernels.

Support Vector Machine (SVM) is a kernel based technique that represents one of the major developments in machine learning algorithm. SVM is a group of supervised learning methods that can be applied for classification and regression. It learns by example to assign labels to objects. SVM has shown its capacities in pattern recognition and a better performance in many domains compared with other machine learning techniques. SVMs have also been successfully applied to an increasingly wide variety of biological applications. Based on empirical results and several classification applications in automatic classification of medical X-ray images, SVM has shown a better generalization performance compared with other classification techniques [40-42].

SVM algorithm takes a set of input data points. It then decides that the data point belong to which possible two classes. The aim is to construct a hyperplane or set of hyperplanes in a high or infinite dimensional space that classifies the data more accurately. Therefore, the basic idea is to find a hyperplane that has the greatest distance to the nearest training data points of any class.

A hyperplane can be defined by the following equation:

wx +b=0 (10)

Where x is the data point lying in the hyperplane, w is normal vector to hyperplane and b is the bias. Figure 9 and 10 show the basics of SVM.

biomedres-classification-SVM

Figure 9. The basics of classification by SVM.

biomedres-Proposed-Model

Figure 10. Proposed Model for RF-SVM.

The idea is to separate two classes (red circle and blue circle), each are labeled as 1 and -1 respectively. Those circles which lie on p1 and p2 are support vectors. For all data points from the hyperplane p (wx + b = 0), the distance between origin and the hyperplane p is (|b|||w||). We consider the patterns from the class -1 that satisfy the equality wx + b = -1, and determine the hyperplane p1; the distance between origin and the hyperplane p1 is equal to ||-1-b|||w||.

Similarly, the patterns from the class +1 satisfy the equality wx + b = +1, and determine the hyperplane p2; the distance between origin and the hyperplane p2 is equal to |+1−b|||w||. Of course, hyperplanes P, P1, and P2 are parallel and no training patterns are 44 located between hyperplanes P1 and P2. Based on the above considerations, the margin between hyperplanes P1 and P2 is 2||w||.

In order to use the SVM methodology to handle the classes are not linearly separable, then the input vectors such as low level feature vectors are mapped to higher dimensional feature space H via a nonlinear transformation, Rd→ H. The kernel function (K(xi,xj)) is used then to construct optimal hyperplane in this high dimensional feature space. This kernel function is a products of input vector equationRadial Basis Function (RBF) kernel and Polynomials kernel are the most common mappings:

Polynomials kernel: K xi,xj=(xi.xj+1)q. (q is the degree of Polynomial)

Radial Basis Function (RBF) kernel: equation (σ is the Gussian sigma)

The above description of SVM is designed for binary classification which only deals with two class labels +1 and -1. It can be extended to multi-class classification where each data point x can be mapped to a specific label y from the set of available labels.

In our proposed system the classification has two divisions: i) training stage and ii) testing stage. In the training stage, various features are extracted from the IAN identified and not identified images. In testing stage based on the knowledge base, the classifier classifies the image into IAN- identified and not identified images. Figure 10 shows the Proposed RF-SVM model for feature selection based classification model.

Experimental Results and Discussions

Data description

Here, 140 OPG DRs collected from MSR Tooth Clinic, Madurai which is composed of 62 OPG radiographs for male patients of age of 10 to 75 years old and 38 for female patients of age of 12 to 74 years old. Around 40 OPG radiographs are posterior dental implants for accuracy of implant position and injuries of IAN. The diagnostic decisions of all the 100 patients were made by a radiologist panel. Table 3 show the summary of experimental data of OPG image and Patients Data Summary with Sex and Age Range in Table 4.

Patient Age (Years) Sex Number of Patients Number of DRs Data Dimensions
10-75 Male-62 100 140 1024 × 564 × 24
Female-38

Table 3. Summary of Experimental Data.

Age Range Male Female Total
10 to 20 15 8 23
21 to 35 27 20 47
36 to 50 22 17 39
51 to 65 10 8 18
66 to 75 9 4 13
Total 83 57 140

Table 4. Patients Data Summary with Sex and Age Range.

Performance assessment

In this study, classification accuracy, sensitivity and specification were used as assessment of the proposed RFSVM model. It defined in Equation (11), (12) and (13).

equation (11)

equation (12)

equation (13)

where TP(True Positive) denote the number of correctly classified IAN injured object, TN(True Negative) denote the number of correctly classified healthy object, FP(False Positive) denote the number of normal cases incorrectly classified. IAN denotes injured object and FN (False Negative) denotes the number of irregular objects incorrectly classified as normal object.

Comparison of classification performance

The obtained experimental results from the proposed technique are given in Table 5 and Table 6. The accuracy of SVM classification accuracy is better than Decision Tree (DT) and Multilayer Perceptron (MLP) classifiers. Both training and testing stages of Random Forest (RF) - Support Vector Machine (SVM) get better classification accuracy of 96.4% and 83.58% respectively. The graphical result Figures 11 and 12 showed the comparison of proposed results with existing Results [13,14,39].

Feature Selection
+ Classifier
Classification Accuracy Sensitivity Specificity
PCA - DT [13] 82.3 76.2 85.2
PCA - MLP [14] 85.4 76.6 90.2
PSO - MLP [39] 88.7 85.2 88.5
RF - SVM (Proposed) 96.4 88.2 99.1

Table 5. Comparison of Training results with Different classifiers.

Feature Selection
+ Classifier
Classification Accuracy Sensitivity Specificity
PCA - DT [13] 78.58 68.6 76.8
PCA - MLP [14] 79.29 69.9 83.8
PSO - MLP [39] 80.71 77.6 80.5
RF - SVM (Proposed) 83.58 79.9 92.8

Table 6. Comparison of Testing results with Different classifiers.

biomedres-Performance-analysis-classification

Figure 11. Performance analysis for classification accuracy.

biomedres-Sensitivity-Specificity-training

Figure 12. Performance analysis for Sensitivity and Specificity for training and classification stage.

In the proposed system, experimental results are compared to medical expert results with proposed system results in Table 7.

IAN Identification Medical Experts PCA-DT PCA-MLP PSO-MLP RF-SVM
No. of Images % No. of Images % No. of Images % No. of Images % No. of Images %
Not Identified 12 8.57 8 5.71 7 5 7 5 5 3.57
Partially Identified 25 17.86 22 15.71 22 15.71 20 14.29 18 12.86
Clearly identified 103 73.57 110 78.58 111 79.29 113 80.71 117 83.58

Table 7. Comparison of Different Classifier Results with Medical Experts.

Above the experimental data in table 7, we got the following results.

Not identified in IAN from medical experts 8.57%, but the proposed system has 3.57% only.

Partially identified in IAN, 17.86% for medical experts and 12.86% for the proposed system.

Clearly identified in IAN, 73.57% for medical experts, 83.58% for proposed system.

The total number of partial and clearly identified IAN is 91.43% for medical experts and 96.44% for proposed system. Comparatively, the proposed system clearly identified IAN is 5% more than medical experts. So, the proposed system has improved accuracy of IAN identification then medical experts. Figure 13 shows the graphical representation of medical expert results with proposed system results.

biomedres-Comparison-IAN-Identification

Figure 13. Comparison of IAN Identification.

In Tables 8-12 to be discussed in the age range classification of IAN identification with medical experts, PCA-DT, PCAMLP, PSO-MLP and RF-SVM respectively. Below the age of 15 and more than age of 65, the visibility of IAN is very limited.

Age Range Not Identified Partially Identified Clearly identified
Male Female Male Female Male Female
10 to 20 4 2 7 5 4 1
21 to 35 0 0 2 1 25 19
36 to 50 0 0 1 0 21 17
51 to 65 1 0 1 2 8 6
66 to 75 3 2 5 1 1 1
Total 8 4 16 9 59 44

Table 8. Summary of IAN identification for medical experts.

Age Range Not Identified Partially Identified Clearly identified
Male Female Male Female Male Female
10 to 20 3 1 6 5 6 2
21 to 35 0 0 2 0 25 20
36 to 50 0 0 1 0 21 17
51 to 65 0 0 1 1 9 7
66 to 75 2 2 5 1 2 1
Total 5 3 15 7 63 47

Table 9. Summary of IAN Identification for PCA-DT Classifier.

Age Range Not Identified Partially Identified Clearly identified
Male Female Male Female Male Female
10 to 20 2 1 6 4 7 3
21 to 35 0 0 0 1 27 19
36 to 50 0 0 1 0 21 17
51 to 65 0 0 1 2 9 6
66 to 75 2 2 5 2 1 1
Total 4 3 13 9 65 46

Table 10. Summary of IAN Identification for PCA-MLP Classifier.

Age Range Not Identified Partially Identified Clearly identified
Male Female Male Female Male Female
10 to 20 2 1 5 4 8 3
21 to 35 0 0 1 0 26 20
36 to 50 0 0 1 0 21 17
51 to 65 0 0 1 1 9 7
66 to 75 3 1 6 1 1 1
Total 5 2 14 6 65 48

Table 11. Summary of IAN Identification for PSO-MLP Classifier.

Age Range Not Identified Partially Identified Clearly identified
Male Female Male Female Male Female
10 to 20 2 1 10 3 3 4
21 to 35 0 0 0 0 27 20
36 to 50 0 0 0 0 22 17
51 to 65 0 0 1 0 9 8
66 to 75 1 1 2 2 6 1
Total 3 2 13 5 67 50

Table 12. Summary of IAN Identification for RF-SVM Classifier.

Conclusion

In this paper, unified RF-SVM model based on EZW based texture and shape features are proposed. It was performed on a database consisting 140 DRs. By the aim of improving the classification accuracy, different conventional features are extracted. Then, Random Forest based Support Vector Machine model was used for feature selection and classification process. Experimental results shows that our method is an effective feature selection method obtain higher classification accuracy of other methods. Other performance analysis like sensitivity and specificity of our proposed model perform get better than other methods. The provided results show effectively classifies the IANI identification.

References