When Gradient Descent Meets Derivative-Free Optimization: A Match Made in Black-Box Scenario

Large pre-trained language models (PLMs) have garnered significant attention for their versatility and potential for solving a wide spectrum of natural language processing (NLP) tasks. However, the cost of running these PLMs may be prohibitive. Furthermore, PLMs may not be open-sourced due to commer...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Hauptverfasser: Han, Chengcheng, Cui, Liqing, Zhu, Renyu, Wang, Jianing, Chen, Nuo, Sun, Qiushi, Li, Xiang, Gao, Ming
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext bestellen
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!