Machine Vision Approach for Classification of Rice Varieties Using Texture Features
The main objective of this study was to assess the machine vision (MV) techniques to classify six Asian rice varieties commonly named as Kachi-Kainat, Kachi-Toota, Kainat-Pakki, Super-Basmati-Kachi, Super-Basmati-Pakki, and Super-Maryam-Kainat (A1, A2, A3, A4, A5, and A6), mainly cultivated in Pakis...
Gespeichert in:
Veröffentlicht in: | International journal of food properties 2021-01, Vol.24 (1), p.1615-1630 |
---|---|
Hauptverfasser: | , , , , , , , , |
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
Zusammenfassung: | The main objective of this study was to assess the machine vision (MV) techniques to classify six Asian rice varieties commonly named as Kachi-Kainat, Kachi-Toota, Kainat-Pakki, Super-Basmati-Kachi, Super-Basmati-Pakki, and Super-Maryam-Kainat (A1, A2, A3, A4, A5, and A6), mainly cultivated in Pakistan, China, India, Bangladesh, and neighboring countries. The sample of each selected rice variety contained 1800 grains, giving a total of 10800 (1800 × 6) grain samples. A cell phone camera captured the actual field digital images dataset in an open climate. All the captured images were enhanced and converted into the standard 8-bit gray-scale format. Six radius-based non-overlapping regions of interest (ROI's) were taken on each captured image inducing a total of 3600 (6 × 600) ROI's image dataset. We have extracted Binary (B), Histogram (H), and Texture (T) features from each image. We converted these forty-three features for each image into 154800 (43 × 3600) feature vector (FV) space to discriminate rice varieties. After optimizing the FV, five MV classifiers, namely; LMT Tree (LMT-T), Meta Classifier via Regression (MCR), Meta Bagging (MB), Tree J48 (T-J48), and Meta Attribute Select Classifier (MAS-C), were deployed attaining the classification accuracies as 97.4%, 97.0%, 96.3%, 95.74%, and 95.2%, respectively. The maximum overall accuracy (MOA) observed was 97.4% by LMT-Tree. |
---|---|
ISSN: | 1094-2912 1532-2386 |
DOI: | 10.1080/10942912.2021.1986523 |