Complex-valued Depth-wise Separable Convolutional Neural Network for Automatic Modulation Classification
Automatic Modulation Classification (AMC) is a critical task in industrial cognitive communication systems. Existing state-of-the-art methods, typified by real-valued convolutional neural networks, have introduced innovative solutions for AMC. However, such models viewed the two constituent componen...
Gespeichert in:
Veröffentlicht in: | IEEE transactions on instrumentation and measurement 2023-07, p.1-1 |
---|---|
Hauptverfasser: | , , |
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext bestellen |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
Zusammenfassung: | Automatic Modulation Classification (AMC) is a critical task in industrial cognitive communication systems. Existing state-of-the-art methods, typified by real-valued convolutional neural networks, have introduced innovative solutions for AMC. However, such models viewed the two constituent components of complex-valued modulated signals as discrete real-valued inputs, causing structural phase damage to original signals and reduced interpretability of the model. In this article, a novel end-to-end AMC model called a complex-valued depth-wise separable convolutional neural network (CDSCNN) is proposed, which adopts complex-valued operation units to enable automatic complex-valued feature learning specifically tailored for AMC. Considering the limited hardware resources available in industrial scenarios, complex-valued depth-wise separable convolution (CDSC) is designed to strike a balance between classification accuracy and model complexity. With an overall accuracy of 62.63% on the RadioML2016.10a dataset, CDSCNN outperforms its counterparts by 1% to 11%. After fine-tuning on the RadioML2016.10b dataset, the overall accuracy reaches 63.15%, demonstrating the robust recognition and generalization capability of CDSCNN. Moreover, the CDSCNN exhibits lower model complexity compared to other methods. |
---|---|
ISSN: | 0018-9456 |
DOI: | 10.1109/TIM.2023.3298657 |