A simple and efficient technique for leaf extraction in complex backgrounds of low resolution mobile photographed images

Low resolution mobile photographed images pose a complex set of research challenges as compared to non-mobile captured images, which really is a significant issue these days. For non-mobile captured and high-resolution photos, current plant recognition systems are the best solution providers. This s...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Veröffentlicht in:Journal of intelligent & fuzzy systems 2022-01, Vol.43 (1), p.773
Hauptverfasser: Pushpa, B R, N Shobha Rani
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
Beschreibung
Zusammenfassung:Low resolution mobile photographed images pose a complex set of research challenges as compared to non-mobile captured images, which really is a significant issue these days. For non-mobile captured and high-resolution photos, current plant recognition systems are the best solution providers. This study proposes the identification and extraction of leaf regions from complex backgrounds to meet the automatic recognition needs of a variety of mobile phone users. Additionally multiple factors complicate the leaf region extraction from complex backgrounds such as varying background patterns, clutters, varying leaf shape/size and varying illumination due to volatile weather conditions. In this paper, a simple and efficient method for leaf extraction from complex background of mobile photographed low resolution images is proposed based on color channel thresholding and morphological operations. A self-built database of 5000 mobile photographed images in realistic environments is adapted for experimentations. Experiments were conducted on various resolution categories, and it was discovered that the proposed model has an average dice similarity measure of 99.5 percent for successful extraction of the leaf region in 13MP mobile photographed images. Furthermore, our comparative investigation reveals that the suggested model outperforms both traditional and state-of-the-art techniques.
ISSN:1064-1246
1875-8967
DOI:10.3233/JIFS-212451