Advanced Learning-Based Inter Prediction for Future Video Coding
In the fourth generation Audio Video coding Standard (AVS4), the Inter Prediction Filter (INTERPF) reduces discontinuities between prediction and adjacent reconstructed pixels in inter prediction. The paper proposes a low complexity learning-based inter prediction (LLIP) method to replace the tradit...
Gespeichert in:
Hauptverfasser: | , , , , |
---|---|
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext bestellen |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
Zusammenfassung: | In the fourth generation Audio Video coding Standard (AVS4), the Inter
Prediction Filter (INTERPF) reduces discontinuities between prediction and
adjacent reconstructed pixels in inter prediction. The paper proposes a low
complexity learning-based inter prediction (LLIP) method to replace the
traditional INTERPF. LLIP enhances the filtering process by leveraging a
lightweight neural network model, where parameters can be exported for
efficient inference. Specifically, we extract pixels and coordinates utilized
by the traditional INTERPF to form the training dataset. Subsequently, we
export the weights and biases of the trained neural network model and implement
the inference process without any third-party dependency, enabling seamless
integration into video codec without relying on Libtorch, thus achieving faster
inference speed. Ultimately, we replace the traditional handcraft filtering
parameters in INTERPF with the learned optimal filtering parameters. This
practical solution makes the combination of deep learning encoding tools with
traditional video encoding schemes more efficient. Experimental results show
that our approach achieves 0.01%, 0.31%, and 0.25% coding gain for the Y, U,
and V components under the random access (RA) configuration on average. |
---|---|
DOI: | 10.48550/arxiv.2411.15759 |