Improved pre-training method, electronic equipment and storage medium
The invention discloses an improved pre-training method, an electronic device and a storage medium, the pre-training method is used for a pre-training model, the pre-training model comprises a unit generation module and a backbone network, and the method comprises the steps: carrying out the down-sa...
Gespeichert in:
Hauptverfasser: | , , , , |
---|---|
Format: | Patent |
Sprache: | chi ; eng |
Schlagworte: | |
Online-Zugang: | Volltext bestellen |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
Zusammenfassung: | The invention discloses an improved pre-training method, an electronic device and a storage medium, the pre-training method is used for a pre-training model, the pre-training model comprises a unit generation module and a backbone network, and the method comprises the steps: carrying out the down-sampling of a voice through the backbone network, and obtaining a first voice representation, using a mask to shield part of the first voice representation to obtain second voice representation, and aggregating the second voice representation to obtain output voice representation; and for the shielded part of the output voice representation, calculating loss with the discrete target extracted by the unit generation module, and performing gradient back propagation in the backbone network. According to the embodiment of the invention, a framework for improving self-supervised speech representation learning through an unsupervised algorithm is provided, the training target of self-supervised learning is optimized, the e |
---|