# Introduction he theory of support vector machines (SVMs), which is based on the idea of structural risk minimization (SRM), is a new classification technique and has drawn much attention on this topic in recent years (Burges, 1998;Cortes and Vapnik 1995;Vapnik, 1995Vapnik, , 1998)). The good generalization ability of SVMs is achieved by finding a large margin between two classes (Bartlett and Shawe-Taylor, 1998; Shawe-Taylor and Bartlett, 1998). In many applications, the theory of SVMs has been shown to provide higher performance than traditional learning machines (Burges, 1998) and has been introduced as powerful tools for solving classification problems. Since the optimal hyper-plane obtained by the SVM depends on only a small part of the data points, it may become sensitive to noises or outliers in the training set (Boser et al., 1992;Zhang, 1999). To solve this problem, one approach is to do some preprocessing on training data to remove noises or outliers, and then use the remaining set learn the decision function (Cao et al., 2003).This method is hard to implement if we do not have enough knowledge about noises or outliers. In many real world applications, we are given a set of training data without knowledge Author : Department of Computer Science and Engineering, Indian School of Mines, Dhanbad, Jharkhand-826004, India. E-mail : vrush2elu@ismu.ac.in Author : Institute of Mathematics & Applications Bhubaneswar, India. E-mail : sarimif@gmail.com about noises or outliers. There are some risks to remove the meaningful data points as noises or outliers. Support Vector Machines have gained much attention in recent years due to their better predictability and ability to theoretically project any data to infinite dimension. It works on the simple basis of separating classes using a hyper plane. In present scenario, any classification method has to deal with thousands of genes provided by micro array data. This is real test for and classification method. Neural networks have shown high potential in dealing with huge amount of data but it cannot overcome redundancy problem. Many highly correlated genes play similar role in classification while many of them could be omitted. Support Vector Machines also could not account for this problem in current form. In SVM, the weights of any two highly correlated features will be quite near and thus both can play significant role in classification. It is a major hindrance in feature selection. In this paper, we present two different approaches for improvement of classification accuracy for linear SVMs. In the first method, redundancy control has been targeted for improve the classification rate. For checking a control on redundancy an matrix 'A' has been introduced in the optimization problem. This matrix keeps a check on weight of a feature according to its correlation with other features. It will be discussed in details in later section of paper. The second method is also an approach to improve the classification performance of linear SVM. It is based on adjustment of bias value in SVM. The results have encouraged us for further probe. In the paper, Section 2 describes the architecture of normal Support vector machine. Section 3 compares the architecture of normal SVM and modified SVM for controlling the redundancy. Section 4 describes the other method for improving the classification accuracy called Orthogonal Vertical Permutator. Section 5 and 6 discusses experimentation and the results obtained respectively. data perfectly into two classes. Since the example data is often not linearly separable, SVM introduce a "kernel induced feature space" which casts data into a higher dimensional space where data is separable. SVM plays a major role in eliminating computational complexity and over fitting (Crammer, Koby, 2001, Drucker, Harris, 1996, Ferris, Michael C, 2002 and T. S. Furey et al. 2000). We are given l training samples {xi, yi}, i = 1,?., l , where each sample has d inputs , and a class label with one of two values (yi?{-1,1}).Now, all hyper-planes in are parameterized by a vector (w),and a constant (b), expressed in the equation ?? .?? + ?? = 0 where w is orthogonal to the hyper-plane. Given such a hyper-plane (w, b) that separates the data, this gives the function ð??"ð??"(??) = ????ð??"ð??"??(??.?? + ??) which correctly systemizes the training. However, a given hyper-plane represented by (w, b) is equally expressed by all pairs {????,?? ??} for ?????+. So we define the canonical hyper-plane to be that which separates the data from the hyper-plane by a distance of at least 1. That is, we consider those that satisfy: Hence, the hyper-plane that optimally separates the data is the one that minimizes This is solved by using Lagrange's multipliers. Where ?? are the Lagrange multipliers.The Lagrangian has to be minimised with respect to ??, ?? and minimised with respect to ???0. Classical Lagrangian duality enables the primal problem, The minimum with respect to ?? and ?? of the Lagrangian, ?? , is given by, Hence, the dual problem is And hence the solution to the problem is given by With constraints, This equation is can be represented as a quadratic form. # iii. Orthogonal Vertical Permutator Orthogonal Vertical Permutator is a reformation of SVM. In OVP, we vary the bias value of SVM which results in vertical permutations of the hyper-plane resulting from SVM. This section of paper focuses on the bias value 'b' in SVM framework. Its theoretical inspiration is being discussed in following section. Bias is the constant term which is added in decision making equation. Since, SVM generates the hyper-plane with the best possible slope, here we have adjusted the bias value to shift this plane using minimization of classification error of both classes. Therefore it can be seen that classification error is less in the later case as compared to the normal SVM. For realizing this plane, an approach similar to gradient descent is used. Bias value is changed by a fraction of its current value depending on the minimization of error. The error in this case is defined in a different way than in usual case. Normally error is defined as, But in this case, we have defined Error as Both error rates are quite different. In first case the, each error has absolute importance and is equally important. But in the other case, the error rate for each class is different and its importance is related to the number of samples in its class. In second case, one class may be classified to very high accuracy at the expense of the other. It leads to higher probability of accuracy rate for one class. SVM is used to categorize datasets into binary data. All the hyper-planes separating the data into two groups are orthogonal to vector w. The variation of bias gives rise to various permutations of the hyper-planes along the vertical. Our model gives rise to vertical permutations of the orthogonal hyper-planes and hence the Orthogonal Vertical Permutator is named. # III. # Result and Analysis This section gives the comparison of percentage accuracy of SVM with modified SVM in Table-I against the sigma values of A matrix. It can be inferred that the modification offers a better accuracy over SVM. The second table depicts a comparison of percentage of accuracy of SVM and OVP-SVM. The percentage accuracy with shifted bias value is better than normal bias value. Figure 2 and figure 3 gives the graphical representation of application of SVM and OVP-SVM on concentric circle dataset and spiral dataset respectively. The comparison analysis of both classification methods is done on the benchmark datasets. Each dataset is validated using double cross fold approach. Linear SVM is used for classification. Therefore only one parameter needs to be tuned i.e. the 'C' which accounts for soft margin classification. If training data was not provided separately then data was analyzed using 5 fold double cross validation. The data was divided into four parts of training data and one part of testing data. This training data was again five folded with four folds for actual training and one fold for parameter adjustment. All the datasets are available at UCI Machine Learning repository [6]. Table 1 shows the effect of change in matrix 'A' on sonar dataset. In rest of cases results are obtained by creating matrix 'A' is made by summing the correlation coefficient of a feature with rest of the features. The experimentation of SVM with changed value of bias is performed on two datasets. The result of concentric circle dataset and of Spiral Dataset is shown in table 2. ![a) SVM Support Vector Machines (SVM's) are learning methods used for binary classification of data. The basic idea is to find a hyper-plane separating n-dimensional T © 2013 Global Journals Inc. (US)](image-2.png "") 21![xi .w + b > +1whenyi = +1 and ????.?? + ?? 1 ???. We can frame this as an optimization problem as: Minimize in (??,??): ||??|| subject to (ð??"ð??"???? ?????? ??=1,?.,??)????(??.???????)?1 b) Modified SVM Before we start the modification over the existing SVM let us understand the method of generating the matrix A.i. Generation of 'A' Matrix 1. On basis of property of features like correlation or mutual information. 2. Using a function of importance or unique data points. 3. We may also use something like gradient descent method.Consider the optimization problem in SVM. We introduce a matrix 'A' of order nxn where 'n' is number of features. If we minimize this optimization problem, the weight vector obtained is different from normal SVMIntroducing Lagrange's multiplier and converting to dual form ii. Comparing Architecture of SVM with Modified SVM The layout of normal SVM has been shown below. A separating hyper-plane in canonical form must satisy the following constraints, ????[+??]?1 ,??=1,?,??. The distance d(w, b; x) of a point x from the hyper-plane (w, b) is © 2013 Global Journals Inc. (US) 12 Year 013 Modification of Support Vector Machine for Microarray Data Analysis If ? ? = ? ? 1+? ? ; Then ? = (?????????? ?????? ? ?????? ??????) 2 ; So, ? = (? ? ? ? + ? ? ? ? ) . ? ? (? ? ? ? + ?) ? 1 ? ? ? where? = ( [? ? (? ? ? ? + ?) ? 1 + ? ? ] ? ? + ? ? = ? ?? ?? = 0 ? ? = ? ?1 ? ? ? ? ? ? ? = ? ? ? ? ? ? ? 1 + ? ? ? = ? ? ? ? ? ? ? ? ? = ? ? ? ? ? ? 1 + ? ? ? ? ?(?, ?; ?) = | < ?, ? ? > +?| ||?|| ?(?, ?) = min ? ? ;? ? =?1 ?(?, ?; ? ? ) + min ? ? ;? ? =1 ?(?, ?; ? ? ) ?(?, ?) = min ? ? ;? ? =?? ? ;? ? =?1 | < ?, ? ? > +?| + min ? ? ;? ? =1 | < ?, ? ? > +?|) ?(?, ?) we can keep a check on redundancy depending on 'A' matrix.](image-3.png "2 1 |") 1![Figure 1 : Depicting the concept of OVP in SVM](image-4.png "Figure 1 :") 23![Figure 2 : Results of SVM and OVP-SVM on Concentric Circle Dataset](image-5.png "Figure 2 :Figure 3 :") 1Modified SVM 2 © 2013 Global Journals Inc. (US) Modification of Support Vector Machine for Microarray Data Analysis The results obtained in both methods outperform the normal SVM and have different advantages. Modified SVM is immune to redundancy and OVP helps in improvising the classification accuracy of SVM and can be beneficial in multi class datasets. The processing was done on Matlab R2009. * On certain integrals of Lipschitz-Hankel type involving products of Bessel functions GEason BNoble INSneddon Phil. Trans. Roy. Soc. London 247 April 1955 * Oxford: Clarendon, 1892 J. ClerkMaxwell A Treatise on Electricity and Magnetism 2 3rd ed. * Fine particles, thin films and exchange anisotropy ISJacobs CPBean Magnetism GTIii HRado Suhl New York Academic 1963 * On the Algorithmic Implementation of Multiclass Kernelbased Vector Machines Koby;Crammer Singer ;RYoram Nicole J. of Machine Learning Research 2 2001 * Support Vector Regression Machines Harris;Drucker Burges JCChristopher Linda;Kaufman AlexanderJSmola VladimirNVapnik Advances in Neural Information Processing Systems 9 MIT Press 1997. 1996 * Interior-point methods for massive support vector machines MichaelCFerris ToddSMunson SIAM Journal on 13 3 2002 * support vector machine classfication and validation of cancer tisue samples using microarray expressino data TSFurey Bioinformatics 16 10 2000 * Electron spectroscopy studies on magneto-optical media and plastic substrate interface YYorozu MHirano KOka YTagawa IEEE Transl * JMagn Japan Digests 9th Annual Conf. Magnetics Japan August 1987. 1982 2 301 * The Technical Writer's Handbook. Mill Valley MYoung 1989 CA: University Science