Low Computational Complexity Digital Predistortion Based on Direct Learning With Covariance Matrix

This paper proposes a novel approach for digital predistortion, based on direct learning architecture (DLA), to reduce the computational complexity. In power amplifier (PA) behavioral models, the coefficients of a Volterra polynomial or a simplified Volterra polynomial are extracted by calculating t...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Veröffentlicht in:IEEE transactions on microwave theory and techniques 2017-11, Vol.65 (11), p.4274-4284
Hauptverfasser: Zonghao Wang, Wenhua Chen, Gongzhe Su, Ghannouchi, Fadhel M., Zhenghe Feng, Yuanan Liu
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext bestellen
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
Beschreibung
Zusammenfassung:This paper proposes a novel approach for digital predistortion, based on direct learning architecture (DLA), to reduce the computational complexity. In power amplifier (PA) behavioral models, the coefficients of a Volterra polynomial or a simplified Volterra polynomial are extracted by calculating the inverse of a time-varying matrix, which is resource-consuming, time-consuming, and power-consuming due to its matrix dimension and inverse operation in a field-programmable gate array. To speed up the computation and save hardware resources, we propose a low computational complexity DLA with covariance matrix that uses the constant covariance matrix to replace the time-varying input signal filled matrix based on a stationary ergodic random process. To verify the proposed method, it was applied to a wideband Doherty gallium nitride (GaN) PA at 2.6 GHz with a 40-MHz orthogonal frequency division multiplexing signal, and to a dual-band Doherty GaN PA at 1.9 and 2.6 GHz with two 20-MHz long-term evolution signals. Experimental results show that the proposed algorithm achieves almost the same performance as the traditional approach, with less than one fifth of the computational quantity.
ISSN:0018-9480
1557-9670
DOI:10.1109/TMTT.2017.2690290