Maxi-Min discriminant analysis via online learning

Bo Xu, Kaizhu Huang*, Cheng Lin Liu

*Corresponding author for this work

Research output: Contribution to journalArticlepeer-review

14 Citations (Scopus)

Abstract

Linear Discriminant Analysis (LDA) is an important dimensionality reduction algorithm, but its performance is usually limited on multi-class data. Such limitation is incurred by the fact that LDA actually maximizes the average divergence among classes, whereby similar classes with smaller divergence tend to be merged in the subspace. To address this problem, we propose a novel dimensionality reduction method called Maxi-Min Discriminant Analysis (MMDA). In contrast to the traditional LDA, MMDA attempts to find a low-dimensional subspace by maximizing the minimal (worst-case) divergence among classes. This "minimal" setting overcomes the problem of LDA that tends to merge similar classes with smaller divergence when used for multi-class data. We formulate MMDA as a convex problem and further as a large-margin learning problem. One key contribution is that we design an efficient online learning algorithm to solve the involved problem, making the proposed method applicable to large scale data. Experimental results on various datasets demonstrate the efficiency and the efficacy of our proposed method against five other competitive approaches, and the scalability to the data with thousands of classes.

Original languageEnglish
Pages (from-to)56-64
Number of pages9
JournalNeural Networks
Volume34
DOIs
Publication statusPublished - Oct 2012

Keywords

  • Dimensionality reduction
  • Handwritten Chinese character recognition
  • Linear discriminant analysis
  • Multi-category classification

Fingerprint

Dive into the research topics of 'Maxi-Min discriminant analysis via online learning'. Together they form a unique fingerprint.

Cite this