MnasFPN: Learning Latency-aware Pyramid Architecture for Object Detection on Mobile Devices
Despite the blooming success of architecture search for vision tasks in resource-constrained environments, the design of on-device object detection architectures have mostly been manual. The few automated search efforts are either centered around non-mobile-friendly search spaces or not guided by on...
Gespeichert in:
Hauptverfasser: | , , , , , , |
---|---|
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext bestellen |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
Zusammenfassung: | Despite the blooming success of architecture search for vision tasks in
resource-constrained environments, the design of on-device object detection
architectures have mostly been manual. The few automated search efforts are
either centered around non-mobile-friendly search spaces or not guided by
on-device latency. We propose MnasFPN, a mobile-friendly search space for the
detection head, and combine it with latency-aware architecture search to
produce efficient object detection models. The learned MnasFPN head, when
paired with MobileNetV2 body, outperforms MobileNetV3+SSDLite by 1.8 mAP at
similar latency on Pixel. It is also both 1.0 mAP more accurate and 10% faster
than NAS-FPNLite. Ablation studies show that the majority of the performance
gain comes from innovations in the search space. Further explorations reveal an
interesting coupling between the search space design and the search algorithm,
and that the complexity of MnasFPN search space may be at a local optimum. |
---|---|
DOI: | 10.48550/arxiv.1912.01106 |