# Introduction dataset is class imbalanced if the classification categories are not approximately equally represented. The level of imbalance (ratio of size of the majority class to minority class) can be as huge as 1:99 [1]. It is noteworthy that class imbalance is emerging as an important issue in designing classifiers [2], [3], [4]. Furthermore, the class with the lowest number of instances is usually the class of interest from the point of view of the learning task [5]. This problem is of great interest because it turns up in many real-world classification problems, such as remote-sensing [6], pollution detection [7], risk management [8], fraud detection [9], and especially medical diagnosis [10]- [13]. There exist techniques to develop better performing classifiers with imbalanced datasets, which are generally called Class Imbalance Learning (CIL) methods. These methods can be broadly divided into two categories, namely, external methods and internal methods. External methods involve preprocessing of training datasets in order to make them balanced, while internal methods deal with modifications of the learning algorithms in order to reduce their sensitiveness to class imbalance [14]. The main advantage of external methods as previously pointed out, is that they are independent of the underlying classifier. Whenever a class in a classification task is under represented (i.e., has a lower prior probability) compared to other classes, we consider the data as imbalanced [15], [16]. The main problem in imbalanced data is that the majority classes that are represented by large numbers of patterns rule the classifier decision boundaries at the expense of the minority classes that are represented by small numbers of patterns. This leads to high and low accuracies in classifying the majority and minority classes, respectively, which do not necessarily reflect the true difficulty in classifying these classes. Most common solutions to this problem balance the number of patterns in the minority or majority classes. Resampling techniques can be categorized into three groups. Under-sampling methods, which create a subset of the original data-set by eliminating instances (usually majority class instances); oversampling methods, which create a superset of the original dataset by replicating some instances or creating new instances from existing ones; and finally, hybrids methods that combine both sampling methods. Among these categories, there exist several different proposals; from this point, we only center our attention in those that have been used in under sampling. Either way, balancing the data has been found to alleviate the problem of imbalanced data and enhance accuracy [15], [16], [17]. Data balancing is performed by, e.g., oversampling patterns of minority classes either randomly or from areas close to the decision boundaries. Interestingly, random oversampling is found comparable to more sophisticated oversampling methods [17]. Alternatively, under-sampling is performed on majority classes either randomly or from areas far away from the decision boundaries. We note that random under-sampling may remove significant patterns and random oversampling may lead to over-fitting, so random sampling should be performed with care. We also note that, usually, selective under sampling of majority classes is more accurate than oversampling of minority class. In this paper, we are laying more stress to propose an external class imbalance learning method for solving the class imbalance problem by performing selective under sampling of majority class. This paper is organized as follows. Section II presets the problem of cluster disjuncts. Section III briefly reviews the data balancing problems and its measures and in Section IV, we discuss the proposed method of MAJOR_CD (Majority Under-sampling based on Cluster Disjunct) for class imbalance learning. Section V presents the imbalanced datasets used to validate the proposed method, while In Section VI, we present the experimental setting and In Section VII discuss, in detail, the classification results obtained by the proposed method and compare them with the results obtained by different existing methods and finally, in Section VIII we conclude the paper. # II. # Problem of Cluster Disjunct In Class Imbalance learning, the numbers of instances in the majority class are outnumbered to the number of instances in the minority class. Furthermore, the minority concept may additionally contain a sub concept with limited instances, amounting to diverging degrees of classification difficulty [18][19]. This, in fact, is the result of another form of imbalance, a within-class imbalance, which concerns itself with the distribution of representative data for sub concepts within a class [20][21][22]. The existence of within-class imbalances is closely intertwined with the problem of small disjuncts, which has been shown to greatly depreciate classification performance [20][21][22][23]. Briefly, the problem of small disjuncts can be understood as follows: A classifier will attempt to learn a concept by creating multiple disjunct rules that describe the main concept [18][19], [23]. In the case of homogeneous concepts, the classifier will generally create large disjuncts, i.e., rules that cover a large portion (cluster) of examples pertaining to the main concept. However, in the case of heterogeneous concepts, small disjuncts, i.e., rules that cover a small cluster of examples pertaining to the main concept, arise as a direct result of underrepresented sub concepts [18][19], [23]. Moreover, since classifiers attempt to learn both majority and minority a concept, the problem of small disjuncts is not only restricted to the minority concept. On the contrary, small disjuncts of the majority class can arise from noisy misclassified minority class examples or underrepresented subconcepts. However, because of the vast representation of majority class data, this occurrence is infrequent. A more common scenario is that noise may influence disjuncts in the minority class. In this case, the validity of the clusters corresponding to the small disjuncts becomes an important issue, i.e., whether these examples represent an actual subconcept or are merely attributed to noise. To solve the above problem of cluster disjuncts we propose the method cluster disjunct minority oversampling technique for class imbalance learning. # III. # Literature Review In this section, we first review the major research about clustering in class imbalance learning and explain why we choose under-sampling as our technique in this paper. The different imbalance data learning approaches are as follows: [25] have proposed a method named EPLogCleaner that can filter out plenty of irrelevant items based on the common prefix of their URLs. M.S.B. PhridviRaj et al. [26] have proposed an algorithm for finding frequent patterns from data streams by performs only one time scan of the database initially and uses the information to find frequent patterns using frequent pattern generation tree. Chumphol Bunkhumpornpat et al. [27] have a new over-sampling technique called DBSMOTE is proposed. DBSMOTE technique relies on a density-based notion of clusters and is designed to oversample an arbitrarily shaped cluster discovered by DBSCAN. DBSMOTE generates synthetic instances along a shortest path from each positive instance to a pseudo centroid of a minorityclass cluster. Matías Di Martino et al. [28] have presented a new classifier developed specially for imbalanced problems, where maximum F-measure instead of maximum accuracy guide the classifier design. V. Garcia et al. [29] have investigated the influence of both the imbalance ratio and the classifier on the performance of several resampling strategies to deal with imbalanced data sets. The study focuses on evaluating how learning is affected when different resampling algorithms transform the originally imbalanced data into artificially balanced class distributions. Table 2 presents recent algorithmic advances in class imbalance learning available in the literature. Obviously, there are many other algorithms which are not included in this table. A profound comparison of the above algorithms and many others can be gathered from the references list. María Dolores Pérez-Godoy et al. [30] have proposed CO2RBFN, a evolutionary cooperativecompetitive model for the design of radial-basis function networks which uses both radial-basis function and the evolutionary cooperative-competitive technique on imbalanced domains. CO2RBFN follows the evolutionary cooperative-competitive strategy, where each individual of the population represents an RBF (Gaussian function will be considered as RBF) and the entire population is responsible for the definite solution. This paradigm provides a framework where an individual of the population represents only a part of the solution, competing to survive (since it will be eliminated if its performance is poor) but at the same time cooperating in order to build the whole RBFN, which adequately represents the knowledge about the problem and achieves good generalization for new patterns. ---------------------------------------------------------------------------RUSBoost A new hybrid sampling/boosting [29] Algorithm. # CO2RBFN A evolutionary cooperative-competitive [30] model for the design of radial-basis function networks which uses both radial-basis function and the evolutionary cooperative-competitive technique. # Improved Adapt the 2-tuples based genetic tuning [33] FRBCSs approach to classification problems showing the good synergy between this method and some FRBCSs. # BSVMs A model assessment of the interplay [37] between various classification decisions using probability, corresponding decision costs, and quadratic program of optimal margin classifier. Der-Chiang Li et al. [31] have suggested a strategy which over-samples the minority class and under-samples the majority one to balance the datasets. For the majority class, they build up the Gaussian type fuzzy membership function and a-cut to reduce the data size; for the minority class, they used the mega-trend diffusion membership function to generate virtual samples for the class. Furthermore, after balancing the data size of classes, they extended the data attribute dimension into a higher dimension space using classification related information to enhance the classification accuracy. Enhong Che et al. [32] have described a unique approach to improve text categorization under class imbalance by exploiting the semantic context in text documents. Specifically, they generate new samples of rare classes (categories with relatively small amount of training data) by using global semantic information of # Global Journal of Computer Science and Technology Volume XIV Issue VII Version I classes represented by probabilistic topic models. In this way, the numbers of samples in different categories can become more balanced and the performance of text categorization can be improved using this transformed data set. Indeed, this method is different from traditional re-sampling methods, which try to balance the number of documents in different classes by re-sampling the documents in rare classes. Such re-sampling methods can cause overfitting. Another benefit of this approach is the effective handling of noisy samples. Since all the new samples are generated by topic models, the impact of noisy samples is dramatically reduced. Alberto Fernández et al. [33] have proposed an improved version of fuzzy rule based classification systems (FRBCSs) in the framework of imbalanced data-sets by means of a tuning step. Specifically, they adapt the 2-tuples based genetic tuning approach to classification problems showing the good synergy between this method and some FRBCSs. The proposed algorithm uses two learning methods in order to generate the RB for the FRBCS. The first one is the method proposed in [34], that they have named the Chi et al.'s rule generation. The second approach is defined by Ishibuchi and Yamamoto in [35] and it consists of a Fuzzy Hybrid Genetic Based Machine Learning (FH-GBML) algorithm. J. Burez et al. [36] have investigated how they can better handle class imbalance in churn prediction. Using more appropriate evaluation metrics (AUC, lift), they investigated the increase in performance of sampling (both random and advanced under-sampling) and two specific modeling techniques (gradient boosting and weighted random forests) compared to some standard modeling techniques. They have advised weighted random forests, as a cost-sensitive learner, performs significantly better compared to random forests. Che-Chang Hsu et al. [37] have proposed a method with a model assessment of the interplay between various classification decisions using probability, corresponding decis ion costs, and quadratic program of optimal margin classifier called: Bayesian Support Vector Machines (BSVMs) learning strategy. The purpose of their learning method is to lead an attractive pragmatic expansion scheme of the Bayesian approach to assess how well it is aligned with the class imbalance problem. In the framework, they did modify in the objects and conditions of primal problem to reproduce an appropriate learning rule for an observation sample. In [38] Alberto Fernández et al. have proposed to work with fuzzy rule based classification systems using a preprocessing step in order to deal with the class imbalance. Their aim is to analyze the behavior of fuzzy rule based classification systems in the framework of imbalanced data-sets by means of the application of an adaptive inference system with parametric conjunction operators. Jordan M. Malof et al. [39] have empirically investigates how class imbalance in the available set of training cases can impact the performance of the resulting classifier as well as properties of the selected set. In this K-Nearest Neighbor (k-NN) classifier is used which is a well-known classifier and has been used in numerous case-based classification studies of imbalance datasets. The bottom line is that when studying problems with imbalanced data, using the classifiers produced by standard machine learning algorithms without adjusting the output threshold may well be a critical mistake. This skewness towards minority class (positive) generally causes the generation of a high number of falsenegative predictions, which lower the model's performance on the positive class compared with the performance on the negative (majority) class. IV. # Methodology In this section, we follow a design decomposition approach to systematically analyze the different imbalanced domains. We first briefly introduce the framework design for our proposed algorithm. The working style of under-sampling tries to remove selective majority instances. Before performing selective under-sampling on the majority subset, the main cluster disjuncts has to be identified and the borderline and noise instances around the cluster disjuncts are to be removed. The number of instances eliminated will belong to the 'k' cluster disjuncts selected by visualization technique. The remaining cluster disjunct instances of the majority subset have to combined with minority set to form improved dataset. Credit History (d). Housing The algorithm 1: MAJOR_CD can be explained as follows, The inputs to the algorithm are majority subclass "p" and minority class "n" with the number of features j. The output of the algorithm will be the average measures such as AUC, Precision, F-measure, TP rate and TN rate produced by the MAJOR_CD methods. The algorithm begins with initialization of k=1 and j=1, where j is the number of cluster disjuncts identified by applying visualization technique on the subset "n" and k is the variable used for looping of j cluster disjuncts. The 'j' value will change from one dataset to other, and depending upon the unique properties of the dataset the value of k can be equal to one also i.e no cluster disjunct attributes can be identified after applying visualization technique on the dataset. In another case attributes related cluster disjunct oversampling can also be performed to improve the skewed dataset. In any case depending on the amount of minority examples generated, the final "strong set" can or cannot be balanced i;e number of majority instances and minority instances in the strong set will or will not be equal. The presented MAJOR_CD algorithm is summarized as below. The datasets is partitioned into majority and minority subsets. As we are concentrating over sampling, we will take minority data subset for further visualization analysis to identify cluster disjuncts. # b) Improve cluster disjunct by removing noisy and borderline instances Minority subset can be further analyzed to find the noisy or borderline instances so that we can eliminate those. For finding the weak instances one of the ways is that find most influencing attributes or features and then remove ranges of the noisy or weak attributes relating to that feature. How to choose the noisy instances relating to that cluster disjunct from the dataset set? We can find a range where the number of samples are less can give you a simple hint that those instances coming in that range or very rare or noise. We will intelligently detect and remove those instances which are in narrow ranges of that particular cluster disjunct. This process can be applied on all the cluster disjuncts identified for each dataset. # c) Forming the strong dataset The minority subset and majority subset is combined to form a strong and balance dataset, which is used for learning of a base algorithm. In this case we have used C4.5 or Naïve Bayes as the base algorithm. V. # Evaluation Metrics To assess the classification results we count the number of true positive (TP), true negative (TN), false positive (FP) (actually negative, but classified as positive) and false negative (FN) (actually positive, but classified as negative) examples. It is now well known that error rate is not an appropriate evaluation criterion when there is class imbalance or unequal costs. In this paper, we use AUC, Precision, F-measure, TP Rate and TN Rate as performance evaluation measures. Let us define a few well known and widely used measures: The Area under Curve (AUC) measure is computed by equation ( 1), (1) The Precision measure is computed by equation( 2), The F-measure Value is computed by equation( 3), # Experimental Framework In this study MAJOR_CD are applied to twelve binary data sets from the UCI repository [40] with different imbalance ratio (IR). Table 3 summarizes the data selected in this study and shows, for each data set, the number of examples (#Ex.), number of attributes (#Atts.), class name of each class (minority and majority) and IR. In order to estimate different measure (AUC, precision, Fmeasure, TP rate and TN rate) we use a tenfold cross validation approach, that is ten partitions for training and test sets, 90% for training and 10% for testing, where the ten test partitions form the whole set. For each data set we consider the average results of the ten partitions. # Table 3 : Summary of benchmark imbalanced datasets To validate the proposed MAJOR_CD algorithm, we compared it with the traditional Support Vector Machines (SVM), C4.5, Functional Trees (FT), SMOTE (Synthetic Minority Oversampling TEchnique) and CART algorithm. # VII. # Results For all experiments, we use existing prototype's present in Weka [41]. We compare the following domain adaptation methods: The True Negative Rate measure is computed by equation ( 5), The True Positive Rate measure is computed by equation ( 4 We compared proposed method MAJOR_CD with the SVM, C4. 5 [42], FT, SMOTE [43] and CART state-of -the-art learning algorithms. In all the datasets using proposed MAJOR_CD learning algorithm. Second, we compare the classification performance of our proposed MAJOR_CD algorithm with the traditional and class imbalance learning methods based on all datasets. Following, we analyze the performance of the method considering the entire original algorithms, without pre-processing, data sets for SVM, C4.5, FT and CART. we also analyze a pre-processing method SMOTE for performance evaluation of MAJOR_CD. The complete table of results for all the algorithms used in this study is shown in Table 4 to 9, where the reader can observe the full test results, of performance of each approach with their associated standard deviation. We Table 4, 5, 6, 7, 8 and 9 reports the results of AUC, Precision, F-measure, TP Rate, TN Rate and 9 provide both the numerical average performance (Mean) and the standard deviation (SD) results. If the proposed technique is better than the compared technique then '?' symbol appears in the column. If the proposed technique is not better than the compared technique then '?' symbol appears in the column. The mean performances were significantly different according to the T-test at the 95% confidence level. The results in the tables show that MAJOR_CD has given a good improvement on all the measures of class imbalance learning. This level of analysis is enough for overall projection of advantages and disadvantages of MAJOR_CD. A two-tailed corrected resampled paired t test is used in this paper to determine whether the results of the cross-validation show that there is a difference between the two algorithms is significant or not. Difference in accuracy is considered significant when the p-value is less. algorithm. The method achieves competitive or better results compared to state-of-the-art baselines. We emphasize that our approach is learnerindependent: visualization can be used in conjunction with many of the existing algorithms in the literature. Furthermore, the fact that we select samples in the model space, as opposed to the feature space, is novel and sets it apart from many previous approaches to transfer learning (for both classification and ranking). This allows us to capture the ''functional change'' assumption and incorporate labeled information in the transfer learning process. Finally, we can say that MAJOR_CD are one of the best alternatives to handle class imbalance problems effectively. This experimental study supports the conclusion that a cluster disjunct approach for cluster detections and elimination can improve the class imbalance learning behavior when dealing with imbalanced data-sets, as it has helped the MAJOR_CD method to be the best performing algorithms when compared with four classical and well-known algorithms: SVM, C4.5, FT and CART and a wellestablished pre-processing technique SMOTE. # VIII. # Conclusion Class imbalance problem have given a scope for a new paradigm of algorithms in data mining. The traditional and benchmark algorithms are worthwhile for discovering hidden knowledge from the data sources, meanwhile class imbalance learning methods can improve the results which are very much critical in real world applications. In this paper we present the class imbalance problem paradigm, which exploits the cluster disjunct concept in the supervised learning research area, and implement it with C4.5 as its base learners. Experimental results show that MAJOR_CD have performed well in the case of multi class imbalance datasets. Furthermore, MAJOR_CD is much less volatile than C4.5. In our future work, we will apply MAJOR_CD to more learning tasks, especially high dimensional feature learning tasks. Another variation of our approach in future work is to analyze the influence of different base classifier effect on the quality of synthetic minority instances generated. # Global Journal of Computer Science and Technology Volume XIV Issue VII Version I 1![Figure 1 : Before (a). Checking Status (b). Duration (c).Credit History (d). Housing](image-2.png "Figure 1 :") 2![Figure 2 : After Applying MAJOR_CD: (a). Checking Status (b). Duration (c). Credit History (d). Housing](image-3.png "Figure 2 :") 1![MAJOR_CD Input: A set of major subclass examples P, a set of minor subclass examples N, jPj < jNj, and Fj, the feature set, j > 0. Output: Average Measure { AUC, Precision, F-Measure, TP Rate, TN Rate} Phase I: Initial Phase: 1: begin 2: k ? 1,j?1. 3: Apply Visualization Technique on subset P, 4: Identify cluster disjunct Cj from P, j= number of cluster disjunct identified in visualization Phase II: Under sampling Phase 5: Apply Oversampling on Cj cluster disjunct from P, 6: repeat 7: k=k+1 8: Remove 'Cj × s' noisy, borderline instances from the majority examples in each cluster disjunct Cj. 9: Until k = j Phase III: Validating Phase 10: Train and Learn A Base Classifier (C4.5) using Improved P and N 11: end The different components of our new proposed framework are elaborated in the next subsections Global Journal of Computer Science and Technology Volume XIV Issue VII Version I Journals Inc. (US) An under-Sampled Approach for Handling Skewed Data Distribution using Cluster Disjuncts](image-4.png "Algorithm 1 :") ![Journals Inc. (US) An under-Sampled Approach for Handling Skewed Data Distribution using Cluster Disjuncts a) Preparation of the Majority and Minority subsets](image-5.png "") 1? SAMPLING METHODS? BASIC SAMPLING METHODS? Under-Sampling? Over-Sampling? ADVANCED SAMPLING METHODS? Tomek Link? The SMOTE approach? Borderline-SMOTE? One-Sided Selection OSS? Neighbourhood Cleaning Rule (NCL)? Bootstrap-based Over-sampling(BootOS)? ENSEMBLE LEARNING METHODS? BAGGING? Asymmetric bagging, SMOTE Bagging? Over Bagging, Under Bagging? Roughly balanced bagging? Lazy Bagging? Random features selection? BOOSTING? Adaboost? SMOTEBoost? DataBoost-IM? RANDOM FORESTS? Balanced Random Forest BRF? Weighted Random Forest WRF? COST-SENSITIVE LEARNING? Direct cost-sensitive learning methods? Methods for cost-sensitive meta-learning? Cost-sensitive meta-learning? Thresholding methods 2ALGORITHM _____________________________________________ DESCRIPTION REFERENECEDCEIDCombining ensemble learning[27]with cost-sensitive learning. 4DatasetsSVMC4.5FTSMOTECARTMAJOR_CD 5An under-Sampled Approach for Handling Skewed Data Distribution using Cluster Disjunctsexperiments we estimate AUC, Precision, F-measure, TPrate and TN rate using 10-fold cross-validation. Weexperimented with 12 standard datasets for UCIaccuracy respectively for fifteen UCI datasets. Tables 4-repository; these datasets are standard benchmarksused in the context of high-dimensional imbalancelearning. Experiments on these datasets have 2 goals.First, we study the class imbalance properties of theYear 20147_______________________________________________________________________________________________________ _______________________________________________________________________________________________________ must emphasize the good results achieved by MAJOR_CD, as it obtains the highest value among all algorithms.Volume XIV Issue VII Version IBreast67.21±7.28?74.28±6.05?68.58±7.52?69.83±7.77?70.22±5.19?72.42±6.32D D D D D D D D ) cDatasets Breast_w Colic Credit-g Diabetes Hepatitis Ionosphere Kv-rs-kp Labor Mushroom Sick Sonar _______________________________________________________________________________________________________ SVM C4.5 FT SMOTE CART MAJOR_CD 96.75±2.00? 95.01±2.73? 95.45±2.52? 96.16±2.06? 94.74±2.60 94.61±2.39 79.78±6.57? 85.16±5.91 79.11± 6.51? 88.53±4.10? 85.37±5.41 85.00±5.97 68.91±4.46? 71.25±3.17? 71.88±3.68? 76.50±3.38? 73.43±4.00? 70.39±4.19 76.55±4.67? 74.49±5.27? 70.62± 4.67? 76.08±4.04? 74.56±5.01? 73.45±5.07 81.90±8.38? 79.22±9.57? 81.40±8.55? 78.35±9.09? 77.10±7.12? 75.29(8.95) 90.26±4.97? 89.74±4.38? 87.10±5.12? 90.28±4.73? 88.87±4.84 88.70(5.31) 99.02±0.54 99.44±0.37 90.61±1.65? 99.66±0.27 99.35±0.43 99.41(0.49) 92.40±11.07? 78.60±16.58? 84.30±16.24? 80.27±11.94 80.03±16.67 80.60(17.16) 100.0±0.00 100.0±0.00 100.0±0.000 100.0±0.00 99.95±0.09 100.00( 0.00) 99.26±0.04? 98.72±0.55? 96.10±0.92? 97.61±0.68? 98.85±0.54 98.68( 0.55) 75.46±9.92? 73.61±9.34? 86.17±8.45? 82.42±7.25? 70.72±9.43? 71.70( 9.00) _______________________________________________________________________________________________________ _______________________________________________________________________________________________________ Breast 0.586±0.102? 0.606±0.087? 0.604±0.082? 0.717±0.084? 0.587±0.110? 0.611±0.095 Breast_w 0.977±0.017? 0.957±0.034? 0.949±0.030? 0.967±0.025? 0.950±0.032? 0.954±0.030( Global Journal of Computer Science and TechnologyColic0.802±0.073? 0.843±0.070?0.777±0.072?0.908±0.040?0.847±0.070?0.850±0.065Credit-g0.650±0.075? 0.647±0.062?0.655±0.044?0.778±0.041?0.716±0.055?0.656±0.065Diabetes0.793±0.072? 0.751±0.0700.668±0.051?0.791±0.041?0.743±0.0710.743±0.067Hepatitis0.757±0.195? 0.668±0.184?0.678±0.139?0.798±0.112?0.563±0.126?0.631(0.182)Ionosphere0.900±0.060? 0.891±0.060?0.831±0.067?0.904±0.053?0.896±0.059?0.885(0.070)Kr-vs-kp0.996±0.005? 0.998±0.0030.906±0.017?0.999±0.0010.997±0.004?0.998(0.002)Labor0.971±0.075? 0.726±0.224?0.844±0.162?0.833±0.127?0.750±0.248?0.802(0.200)© 2014 Global Journals Inc. (US)_______________________________________________________________________________________________________ 6DatasetsSVMC4.5FTSMOTECARTMAJOR_CD 7DatasetsSVMC4.5FTSMOTECARTMAJOR_CD 8_______________________________________________________________________________________________________1.000±0.001.000±0.001.000±0.001.000±0.000.999±0.0011.000±0.00Sick0.990±0.014? 0.952±0.040?0.795±0.053?0.962±0.025?0.954±0.043?0.948(0.042)Sonar _______________________________________________________________________________________________________ 0.771±0.103? 0.753±0.113? 0.859±0.086? 0.814±0.090? 0.721±0.106? 0.725(0.100)______________________________________________________________________________________________________________________________________________________________________________________________________________Breast0.745±0.051?0.753±0.042?0.762±0.051?0.710±0.075?0.728±0.038?0.732±0.043Year 2014Breast_w 0.988±0.019? Colic 0.845±0.060? Credit-g 0.776±0.033? Diabetes 0.793±0.037?0.965±0.026? 0.851±0.055? 0.767±0.025? 0.797±0.045?0.964±0.026? 0.839±0.062? 0.791±0.027? 0.764±0.036?0.974±0.025? 0.853±0.057? 0.768±0.034? 0.781±0.064?0.968±0.026? 0.853±0.053? 0.779±0.030? 0.782±0.0420.961±0.027 0.843±0.061 0.758±0.030 0.782±0.0488Hepatitis Ionosphere 0.906±0.080? 0.895±0.084 0.604±0.271? 0.510±0.371?0.546±0.333? 0.938±0.073?0.709±0.165? 0.934±0.049?0.232±0.334? 0.868±0.096?0.429(0.325) 0.894(0.080)Volume XIV Issue VII Version IKr-vs-kp Labor Mushroom 1.000±0.000 0.991±0.008? 0.915±0.197? Sick 0.997±0.003? Sonar 0.764±0.119? _______________________________________________________________________________________________________ 0.994±0.006 0.905±0.021? 0.996±0.005? 0.993±0.007? 0.994(0.006) 0.696±0.359? 0.802±0.250? 0.871±0.151? 0.715±0.355? 0.738(0.300) 1.000±0.000 1.000±0.000 1.000±0.000 0.999±0.002 1.000±0.000 0.992±0.005 0.975±0.007? 0.983±0.007? 0.992±0.005 0.992(0.005) 0.728±0.121? 0.883±0.100? 0.863±0.068? 0.709±0.118? 0.715(0.108) Breast 0.781±0.059? 0.838±0.040? 0.776±0.057? 0.730±0.076? 0.813±0.038? 0.823±0.043 Breast_w 0.965±0.019? 0.962±0.021? 0.975±0.016? 0.960±0.022? 0.959±0.020 0.958±0.019 _______________________________________________________________________________________________________ _______________________________________________________________________________________________________D D D D ) cColic0.833±0.055?0.888±0.044? 0.838±0.054?0.880±0.042?0.890±0.040?0.883±0.046( Global Journal of Computer Science and TechnologyDatasets Credit-g Diabetes Hepatitis Ionosphere 0.787±0.098? SVM 0.802±0.027 0.778±0.037? 0.469±0.265? Kv-rs-kp 0.911±0.016? Labor 0.794±0.211? Mushroom 1.000±0.000 Sick 0.979±0.005? Sonar 0.844±0.099? _______________________________________________________________________________________________________ C4.5 FT SMOTE CART MAJOR_CD 0.805±0.022? 0.779±0.034? 0.787±0.034? 0.820±0.028? 0.794±0.032 0.806±0.044? 0.827±0.038? 0.741±0.046? 0.812±0.040? 0.794±0.041 0.409±0.272? 0.557±0.207? 0.677±0.138? 0.179±0.235? 0.375(0.258) 0.850±0.066? 0.855±0.079? 0.905±0.048? 0.841±0.070? 0.843(0.078) 0.995±0.004 0.991±0.005? 0.995±0.004 0.994±0.004 0.994(0.005) 0.636±0.312? 0.879±0.195? 0.793±0.132? 0.660±0.316? 0.734(0.280) 1.000±0.000 1.000±0.000 1.000±0.000 0.999±0.001 1.000±0.000 0.993±0.003? 0.996±0.003? 0.987±0.004? 0.994±0.003 0.993(0.003) 0.716±0.105? 0.753±0.102? 0.861±0.061? 0.672±0.106? 0.704(0.105) _______________________________________________________________________________________________________ _______________________________________________________________________________________________________ Breast 0.806±0.091? 0.947±0.060? 0.815±0.095? 0.763±0.117? 0.926±0.081? 0.941±0.061 Breast_w 0.967±0.025? 0.959±0.033? 0.962±0.029? 0.947±0.035? 0.952±0.034? 0.956±0.032 Colic 0.832±0.075? 0.931±0.053? 0.835±0.077? 0.913±0.058? 0.932±0.050 0.931±0.062 Credit-g 0.815±0.041? 0.847±0.036? 0.783±0.052? 0.810±0.058? 0.869±0.047? 0.835±0.055 Diabetes 0.795±0.054? 0.821±0.073? 0.868±0.065? 0.712±0.089? 0.848±0.066? 0.811±0.067Hepatitis0.448±0.273?0.374±0.256?0.573±0.248?0.681±0.188?0.169±0.236?0.371(0.272)Ionosphere 0.689±0.131?0.821±0.107?0.820±0.114?0.881±0.071?0.830±0.112?0.807(0.115)Kv-rs-kp0.916±0.021?0.995±0.0050.990±0.007?0.995±0.0060.995±0.0060.994(0.007)Labor0.845±0.243?0.640±0.349?0.885±0.234?0.765±0.194?0.665±0.359?0.775(0.321)Mushroom 1.000±0.0001.000±0.0001.000±0.0001.000±0.0001.000±0.0001.000±0.000Sick0.984±0.006?0.995±0.0040.995±0.0040.990±0.005?0.996±0.003?0.994(0.004)_______________________________________________________________________________________________________ Sonar 0.820±0.131? 0.721±0.140? 0.757±0.136? 0.865±0.090? 0.652±0.137? 0.708(0.147) 9_______________________________________________________________________________________________________Datasets _______________________________________________________________________________________________________ SVM C4.5 FT SMOTE CART MAJOR_CDBreast0.260±0.1410.335±0.166?0.151±0.164?0.622±0.137?0.173±0.164?0.259±0.134Breast_w0.932±0.052?0.977±0.037?0.931±0.060?0.975±0.024?0.940±0.051?0.928±0.053Colic0.717±0.119? 0.734±0.118?0.731±0.121?0.862±0.063?0.720±0.114?0.727±0.125Credit-g0.398±0.085? 0.469±0.098?0.371±0.105?0.713±0.056?0.421±0.102?0.419±0.092Diabetes0.603±0.111?0.574±0.095?0.567±0.105?0.807±0.077?0.554±0.113?0.601±0.117Hepatitis0.900±0.097?0.882±0.092?0.942±0.093?0.837±0.109?0.928±0.094?0.867(0.100)Ionosphere 0.940±0.055? 0.949±0.046? Kv-rs-kp 0.993±0.007? 0.990±0.009? Labor 0.865±0.197? 0.945±0.131? Mushroom 1.000±0.000 1.000±0.0000.933±0.063? 0.987±0.010? 0.843±0.214? 1.000±0.0000.928±0.057? 0.998±0.003? 0.847±0.187? 1.000±0.0000.921±0.066? 0.992±0.008? 0.877±0.192? 0.999±0.0020.936(0.054) 0.994(0.007) 0.827(0.192) 1.000±0.000Year 2014Sick Sonar _______________________________________________________________________________________________________ 0.875±0.071 0.974±0.026? 0.846±0.080? 0.872±0.053? 0.876±0.078? 0.874(0.074) 0.749±0.134? 0.752±0.148? 0.762±0.145? 0.752±0.113? 0.756±0.121? 0.724(0.122)9than 0.05 (confidence level is greater than 95%). In discussion of results, if one algorithm is stated to be better or worse than another then it is significantly better or worse at the 0.05 level. We can make a global analysis of results combining the results offered by Tables from 4-9: ? Our proposal, MAJOR_CD are the best performing one when the data sets are no preprocessed. outperforms the pre-processing SMOTE methods and this hypothesis is confirmed by including standard deviation variations. We have considered aVolume XIV Issue VII Version Icomplete competitive set of methods and an improvement of results is expected in the( D D D D D D D D ) cbenchmark algorithms i;e SVM, C4.5, FT and CART. However, they are not able to outperform MAJOR_CD. In this sense, the competitive edge of MAJOR_CD can be seen. ? Considering that MAJOR_CD behaves similarly or not effective than SMOTE shows the unique properties of the datasets where there is scope of improvement in minority subset and not in majority subset. Our MAJOR_CD can only consider improvements in majority subset which is not effective for some unique property datasets. The contributions of this work are twofold: A general strategy to handle class imbalance problem: This is scalable, flexible, and modular, allowing the many existing supervised methods to be as a baseGlobal Journal of Computer Science and Technology © 2014 Global Journals Inc. (US)An under-Sampled Approach for Handling Skewed Data Distribution using Cluster Disjuncts © 2014 Global Journals Inc. (US) * Fast asymmetric learning for cascade face detection JWu SCBrubaker MDMullin JMRehg IEEE Trans. Pattern Anal. Mach. Intell 30 3 Mar. 2008 * Mining with rarity: A unifying framework GMWeiss ACM SIGKDD Explor. Newslett 6 1 Jun. 2004 * NVChawla NJapkowicz AKolcz Special Issue Learning Imbalanced Datasets 2004 6 * Ground-level ozone prediction by support vector machine approach with a cost-sensitive classification scheme W.-ZLu DWang Sci. Total. Enviro 395 2-3 2008 * Evaluation of neural networks and data mining methods on a credit assessment task for class imbalance problem Y.-MHuang C.-MHung HCJiau Nonlinear Anal. R. World Appl 7 4 2006 * Combating imbalance in network intrusion datasets DCieslak NChawla AStriegel IEEE Int. Conf. Granular Comput 2006 * Training neural network classifiers for medical decision making: The effects of imbalanced datasets on classification performance MAMazurowski PAHabas JMZurada JYLo JABaker GDTourassi Neural Netw 21 2-3 2008 * Comparison of different strategies of utilizing fuzzy clustering in structure identification AFreitas ACosta-Pereira PBrazdil Data Warehousing Knowl. Discov Lecture Notes Series in Computer Science ISong JEder TNguyen .KEds IBKilic¸ Tu¨rksen 2007 177 Costsensitive decision trees applied to medical data * A methodological approach to the classification of dermoscopy images MECelebi HAKingravi BUddin HIyatomi YAAslandogan WVStoecker RHMoss Comput.Med. Imag. Grap 31 6 2007 * Robust BMPM training based on second-order cone programming and its application in medical diagnosis XPeng IKing Neural Netw 21 2-3 2008. 2007 Springer * FSVM-CIL: Fuzzy Support Vector Machines for Class Imbalance Learning RukshanBatuwita Vasile Palade IEEE TRANSACTIONS ON FUZZY SYSTEMS 18 3 2010. JUNE 2010 * The Class Imbalance Problem: A Systematic Study NJapkowicz SStephen Intelligent Data Analysis 6 2002 * Addressing the Curse of Imbalanced Training Sets: One-Sided Selection MKubat SMatwin Proc. 14th Int'l Conf. Machine Learning 14th Int'l Conf. Machine Learning 1997 * A Study of the Behavior of Several Methods for GE A P ABatista RCPrati MCMonard * NVChawla NJapkowicz AKotcz Proc. ICML Workshop Learn. Imbalanced Data Sets ICML Workshop Learn. Imbalanced Data Sets 2003 * \ Balancing Machine Learning Training Data NJapkowicz Proc. AAAI Workshop Learn. Imbalanced Data Sets AAAI Workshop Learn. Imbalanced Data Sets 2000. 2004.1 6 * Induction of Decision Trees JRQuinlan Machine Learning 1986 1 * Class Imbalances versus Small Disjuncts TJo NJapkowicz ACM SIGKDD Explorations Newsletter 6 1 2004 * Class Imbalances: Are We Focusing on the Right Issue? NJapkowicz Proc. Int'l Conf. Machine Learning, Workshop Learning from Imbalanced Data Sets II Int'l Conf. Machine Learning, Workshop Learning from Imbalanced Data Sets II 2003 * Class Imbalances versus Class Overlapping: An Analysis of a Learning 22. System Behavior RCPrati GE A P ABatista MCMonard Proc. Mexican Int'l Conf. Artificial Intelligence Mexican Int'l Conf. Artificial Intelligence 2004 * Mining with Rarity: A Unifying Framework GMWeiss Procedia -Social and Behavioral Sciences 6 1 2004. 2013 ACM SIGKDD Explorations Newsletter * EPLogCleaner: Improving Data Quality of Enterprise Proxy Logs for Efficient Web Usage Mining HongzhouSha TingwenLiu PengQin YongSun QingyunLiu Procedia Computer Science 17 2013 * Data miningpast, present and future -a typical survey on data Streams MS BPhridviraj CVGururao ; Chidchanok Lursinsap Chumphol Bunkhumpornpat, Krung Sinapiromsaran 2014. 2012 12 Appl Intell * Novel classifier scheme for imbalanced problems Matías DiMartino AliciaFernández PabloIturralde FedericoLecumberry Pattern Recognition Letters 34 2013 * On the effectiveness of preprocessing methods when dealing with different levels of class imbalance VGarcia JSSanchez RAMollineda Knowledge-Based Systems 25 2012 * Analysis of an evolutionary RBFN design algorithm, CO2RBFN, for imbalanced data sets AlbertoMaría Dolores Pérez-Godoy AntonioJesúsFernández María JoséRivera JesusDel Pattern Recognition Letters 31 2010 * Exploiting probabilistic topic models to improve text categorization under class imbalance EnhongChe YanggangLin HuiXiong QimingLuo HaipingMa C4.5: Programs for Machine Learning .J RQuinlan San Mateo, CA Morgan Kaufmann Publishers 1993 1st ed * SMOTE: Synthetic minority over-sampling technique NChawla KBowyer PKegelmeyer J. Artif. Intell. Res 16 2002 * A learning method for the class imbalance problem with medical data sets Der-ChiangLi Chiao-Wenliu Susanc Hu Computers in Biology and Medicine 40 2010 * Information Processing and Management 47 2011 * On the 2-tuples based genetic tuning performance for fuzzy rule based classification systems in imbalanced data-sets AlbertoFernández María José DelJesus FranciscoHerrera Information Sciences 180 2010 * ZChi HYan TPham Fuzzy Algorithms with Applications to Image Processing and Pattern Recognition World Scientific 1996 * Hybridization of fuzzy GBML approaches for pattern classification problems HIshibuchi TYamamoto TNakashima IEEE Transactions on System, Man and Cybernetics B 35 2 2005 * Handling class imbalance in customer churn prediction JBurez DVan Den Poel Expert Systems with Applications 36 2009 * Bayesian decision theory for support vector machines: Imbalance measurement and feature optimization Che-ChangHsu Kuo-ShongWang Shih-HsingChang Expert Systems with Applications 38 2011 * On the influence of an adaptive inference system in fuzzy rule based classification systems for imbalanced data-sets AlbertoFernández María José DelJesus FranciscoHerrera Expert Systems with Applications 36 2009 * The effect of class imbalance on case selection for case-based classifiers: An empirical study in the context of medical decision support JordanMMalof MaciejAMazurowski GeorgiaDTourassi Neural Networks 25 2012 * UCI Repository of Machine Learning Database (School of Information and Computer Science) A AsuncionDNewman 2007 Irvine, CA Univ * Data Mining: Practical machine learning tools and techniques IHWitten EFrank 2005 San Francisco 2nd edition Morgan Kaufmann