Authors
Yuting Yang, Pei Huang, Juan Cao, Jintao Li, Yun Lin, Feifei Ma
Publication date
2024/8
Journal
Frontiers of Computer Science
Volume
18
Issue
4
Pages
184318
Publisher
Higher Education Press
Description
Recent years have seen the wide application of natural language processing (NLP) models in crucial areas such as finance, medical treatment, and news media, raising concerns about the model robustness and vulnerabilities. We find that prompt paradigm can probe special robust defects of pre-trained language models. Malicious prompt texts are first constructed for inputs and a pre-trained language model can generate adversarial examples for victim models via maskfilling. Experimental results show that prompt paradigm can efficiently generate more diverse adversarial examples besides synonym substitution. Then, we propose a novel robust training approach based on prompt paradigm which incorporates prompt texts as the alternatives to adversarial examples and enhances robustness under a lightweight minimax-style optimization framework. Experiments on three real-world tasks and two deep neural …
Total citations
Scholar articles