Survey of Research on Extended Models of Pre-trained Language Models

In recent years,the proposal of Transformer neural network has greatly promoted the development of pre-training technology.At present,pre-training models based on deep learning have become a research hotspot in the field of natural language processing.Since the end of 2018,BERT has achieved optimal...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Veröffentlicht in:Ji suan ji ke xue 2022-01, Vol.49, p.43
Hauptverfasser: Abulizi, Abudukelimu, Zhang, Yu-ning, Yasen, Alimujiang, Guo, Wen-qiang, Halidanmu, Abudukelimu
Format: Artikel
Sprache:chi
Schlagworte:
Online-Zugang:Volltext
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
Beschreibung
Zusammenfassung:In recent years,the proposal of Transformer neural network has greatly promoted the development of pre-training technology.At present,pre-training models based on deep learning have become a research hotspot in the field of natural language processing.Since the end of 2018,BERT has achieved optimal results in multiple natural language processing tasks.A series of improved pre-training models based on BERT have been proposed one after another,and pre-training model extension models designed for various scenarios have also appeared.The expansion of pre-training models from single-language to tasks such as crosslanguage,multi-modality,and light-weighting has enabled natural language processing to enter a new era of pre-training.This paper mainly summarizes the research methods and research conclusions of lightweight pre-training models,knowledge-incorporated pre-training models,cross-modal pre-training language models and cross-language pre-training language models,as well as the main challenges faced by the pre
ISSN:1002-137X