Enhanced hierarchical model of object recognition based on a novel patch selection method in salient regions

The biologically inspired hierarchical model for object recognition, Hierarchical Model and X (HMAX), has attracted considerable attention in recent years. HMAX is robust (i.e. shift- and scale-invariant), but its use of random-patch-selection makes it sensitive to rotational deformation, which heav...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Veröffentlicht in:IET computer vision 2015-10, Vol.9 (5), p.663-672
Hauptverfasser: Lu, Yan-Feng, Kang, Tae-Koo, Zhang, Hua-Zhen, Lim, Myo-Taeg
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext bestellen
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
Beschreibung
Zusammenfassung:The biologically inspired hierarchical model for object recognition, Hierarchical Model and X (HMAX), has attracted considerable attention in recent years. HMAX is robust (i.e. shift- and scale-invariant), but its use of random-patch-selection makes it sensitive to rotational deformation, which heavily limits its performance in object recognition. The main reason is that numerous randomly chosen patches are often orientation selective, thereby leading to mismatch. To address this issue, the authors propose a novel patch selection method for HMAX called saliency and keypoint-based patch selection (SKPS), which is based on a saliency (attention) mechanism and multi-scale keypoints. In contrast to the conventional random-patch-selection-based HMAX model that involves huge amounts of redundant information in feature extraction, the SKPS-based HMAX model (S-HMAX) extracts a very few features while offering promising distinctiveness. To show the effectiveness of S-HMAX, the authors apply it to object categorisation and conduct experiments on the CalTech101, TU Darmstadt, ImageNet and GRAZ01 databases. The experimental results demonstrate that S-HMAX outperforms conventional HMAX and is very comparable with existing architectures that have a similar framework.
ISSN:1751-9632
1751-9640
1751-9640
DOI:10.1049/iet-cvi.2014.0249