Class consistent k-means: Application to face and action recognition

► We introduce a new algorithm, class consistent k-means clustering (CCKM). ► We introduce Hierarchical CCKM to learn a discriminative quantization tree. ► A new multi-class voting-based classification framework is built using CCKM. ► Our approach achieves state of the art performance on several pop...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Veröffentlicht in:Computer vision and image understanding 2012-06, Vol.116 (6), p.730-741
Hauptverfasser: Jiang, Zhuolin, Lin, Zhe, Davis, Larry S.
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
Beschreibung
Zusammenfassung:► We introduce a new algorithm, class consistent k-means clustering (CCKM). ► We introduce Hierarchical CCKM to learn a discriminative quantization tree. ► A new multi-class voting-based classification framework is built using CCKM. ► Our approach achieves state of the art performance on several popular datasets. A class-consistent k-means clustering algorithm (CCKM) and its hierarchical extension (Hierarchical CCKM) are presented for generating discriminative visual words for recognition problems. In addition to using the labels of training data themselves, we associate a class label with each cluster center to enforce discriminability in the resulting visual words. Our algorithms encourage data points from the same class to be assigned to the same visual word, and those from different classes to be assigned to different visual words. More specifically, we introduce a class consistency term in the clustering process which penalizes assignment of data points from different classes to the same cluster. The optimization process is efficient and bounded by the complexity of k-means clustering. A very efficient and discriminative tree classifier can be learned for various recognition tasks via the Hierarchical CCKM. The effectiveness of the proposed algorithms is validated on two public face datasets and four benchmark action datasets.
ISSN:1077-3142
1090-235X
DOI:10.1016/j.cviu.2012.02.004