MarNASNets: Toward CNN Model Architectures Specific to Sensor-Based Human Activity Recognition

Deep learning (DL) models for sensor-based human activity recognition (HAR) are still in their nascent stages compared with image recognition. HAR’s inference is generally implemented on edge devices such as smartphones because of its secure privacy. However, lightweight DL models for HAR, while mee...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Veröffentlicht in:IEEE sensors journal 2023-08, Vol.23 (16), p.18708-18717
Hauptverfasser: Kobayashi, Satoshi, Hasegawa, Tatsuhito, Miyoshi, Takeru, Koshino, Makoto
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
Beschreibung
Zusammenfassung:Deep learning (DL) models for sensor-based human activity recognition (HAR) are still in their nascent stages compared with image recognition. HAR’s inference is generally implemented on edge devices such as smartphones because of its secure privacy. However, lightweight DL models for HAR, while meeting the hardware limitations, are lacking. In this study, using the neural architecture search (NAS), we investigated an effective DL model architecture that can be used for inference on smartphones. We designed multiple search spaces for the type of convolution, the kernel size of the convolution process, the type of skip operation, the number of layers, and the number of output filters by Bayesian optimization. We propose models called mobile-aware convolutional neural network (CNN) for sensor-based HAR by NAS (MarNASNets). We constructed four MarNASNet networks, MarNASNet-A to -D, each with a different model size and a parameter search space of four patterns. Experimental results show that MarNASNets achieve the same accuracy as the existing CNN architectures with fewer parameters and are effective model architectures for on-device and sensor-based HAR. We also developed Activitybench, an iOS app, for measuring model performance on smartphones, and evaluated the on-device performance of each model. MarNASNets’ exploration achieved accuracy comparable to the existing CNN models with smaller model sizes. MarNASNet-C achieved accuracies of 92.60%, 94.52%, and 88.92% for Human Activity Sensing Consortium (HASC), UCI, and Wireless Sensor Data Mining (WISDM), respectively. Especially for HASC and UCI, MarNASNet-C achieved the highest accuracies despite the small model size. Their latency was also comparable to that of the existing CNN models, enabling real-time on-device inference.
ISSN:1530-437X
1558-1748
DOI:10.1109/JSEN.2023.3292380