Please wait a minute...
Front. Inform. Technol. Electron. Eng.  2011, Vol. 12 Issue (5): 362-370    DOI: 10.1631/jzus.C1000361
    
Integrating outlier filtering in large margin training
Xi-chuan Zhou*,1, Hai-bin Shen2, Jie-ping Ye3
1 College of Communication Engineering, Chongqing University, Chongqing 400044, China 2 School of Electrical Engineering, Zhejiang University, Hangzhou 310027, China 3 Department of Computer Science and Engineering, Arizona State University, Tempe 85281, USA
Download:   PDF(433KB)
Export: BibTeX | EndNote (RIS)      

Abstract  Large margin classifiers such as support vector machines (SVM) have been applied successfully in various classification tasks. However, their performance may be significantly degraded in the presence of outliers. In this paper, we propose a robust SVM formulation which is shown to be less sensitive to outliers. The key idea is to employ an adaptively weighted hinge loss that explicitly incorporates outlier filtering in the SVM training, thus performing outlier filtering and classification simultaneously. The resulting robust SVM formulation is non-convex. We first relax it into a semi-definite programming which admits a global solution. To improve the efficiency, an iterative approach is developed. We have performed experiments using both synthetic and real-world data. Results show that the performance of the standard SVM degrades rapidly when more outliers are included, while the proposed robust SVM training is more stable in the presence of outliers.

Key wordsSupport vector machines      Outlier filter      Semi-definite programming      Multi-stage relaxation     
Received: 15 October 2010      Published: 09 May 2011
CLC:  TP301  
Cite this article:

Xi-chuan Zhou, Hai-bin Shen, Jie-ping Ye. Integrating outlier filtering in large margin training. Front. Inform. Technol. Electron. Eng., 2011, 12(5): 362-370.

URL:

http://www.zjujournals.com/xueshu/fitee/10.1631/jzus.C1000361     OR     http://www.zjujournals.com/xueshu/fitee/Y2011/V12/I5/362


Integrating outlier filtering in large margin training

Large margin classifiers such as support vector machines (SVM) have been applied successfully in various classification tasks. However, their performance may be significantly degraded in the presence of outliers. In this paper, we propose a robust SVM formulation which is shown to be less sensitive to outliers. The key idea is to employ an adaptively weighted hinge loss that explicitly incorporates outlier filtering in the SVM training, thus performing outlier filtering and classification simultaneously. The resulting robust SVM formulation is non-convex. We first relax it into a semi-definite programming which admits a global solution. To improve the efficiency, an iterative approach is developed. We have performed experiments using both synthetic and real-world data. Results show that the performance of the standard SVM degrades rapidly when more outliers are included, while the proposed robust SVM training is more stable in the presence of outliers.

关键词: Support vector machines,  Outlier filter,  Semi-definite programming,  Multi-stage relaxation