CSpace  > 中国科学院计算技术研究所期刊论文  > 英文
RobustPrompt: Learning to defend against adversarial attacks with adaptive visual prompts
Liu, Chang1; Xiang, Wenzhao2,3; Dong, Yinpeng6; Zhang, Xingxing4; Wang, Liyuan4; Duan, Ranjie4,5; Zheng, Shibao1; Su, Hang6
2025-04-01
发表期刊PATTERN RECOGNITION LETTERS
ISSN0167-8655
卷号190页码:161-168
摘要Adversarial training stands out as one of the most effective techniques for enhancing robustness by enriching the training data with adversarial examples. Nonetheless, when faced with various perturbation budgets, the model's performance can suffer notable degradation. This occurs because different perturbations induce distinct distribution shifts in adversarial examples. In order to enhance performance on specific perturbations, fine-tuning is commonly employed. However, this approach can lead to catastrophic forgetting, where improvements on specific tasks come at the cost of degrading performance on previously learned ones. We frame this challenge as an incremental domain learning problem in continual learning. Inspired by the application of prompt techniques in vision models, we introduce RobustPrompt, which integrates additional guidance information regarding perturbation characteristics into the adversarial training process. This approach enables the model to adaptively enhance its robustness under varying budget perturbations. Specifically, we define an adaptive prompt pool composed of a noise level predictor and corresponding prompts for different perturbations. During training, prompts are injected into different layers of the model, thereby guiding the model to focus on correct features. Experiments demonstrate that RobustPrompt enhances the adversarial robustness of the state-of-the-art Swin Transformer Base model, achieving an average improvement of 61.1% against PGD attacks and 56.9% against AutoAttack across five white-box settings; an average improvement of 76.1% against VMI-FGSM attacks across five black-box settings; an average improvement of 53.7% on five datasets with natural noise. Our results underscore the potential of RobustPrompt as a useful tool for enhancing the reliability and robustness of transformers in image classification tasks.
关键词Adversarial robustness Prompt tuning Continual learning Adaptive prompt pool
DOI10.1016/j.patrec.2025.02.015
收录类别SCI
语种英语
资助项目NSFC[62076147] ; NSFC[U19A2081] ; NSFC[U19B2034] ; NSFC[U1811461] ; NSFC[62071292] ; NSFC[U21B2013] ; STCSM[22DZ2229005] ; Alibaba Group through Alibaba Innovative Research Program, a grant from Tsinghua Institute for Guo Qiang ; High Performance Computing Center, Tsinghua University
WOS研究方向Computer Science
WOS类目Computer Science, Artificial Intelligence
WOS记录号WOS:001436909400001
出版者ELSEVIER
引用统计
文献类型期刊论文
条目标识符http://119.78.100.204/handle/2XEOYT63/40705
专题中国科学院计算技术研究所期刊论文_英文
通讯作者Zheng, Shibao; Su, Hang
作者单位1.Shanghai Jiao Tong Univ, Inst Image Commun & Networks Engn, Dept Elect Engn EE, Shanghai 200240, Peoples R China
2.PengCheng Lab, Shenzhen 518055, Peoples R China
3.Chinese Acad Sci, Inst Comp Technol, CAS, Key Lab Intelligent Informat Proc, Beijing 100190, Peoples R China
4.Tsinghua Univ, Beijing 100084, Peoples R China
5.Alibaba Grp, Beijing 100102, Peoples R China
6.Tsinghua Univ, Inst AI, Dept Comp Sci & Technol, THBI Lab, Beijing 100084, Peoples R China
推荐引用方式
GB/T 7714
Liu, Chang,Xiang, Wenzhao,Dong, Yinpeng,et al. RobustPrompt: Learning to defend against adversarial attacks with adaptive visual prompts[J]. PATTERN RECOGNITION LETTERS,2025,190:161-168.
APA Liu, Chang.,Xiang, Wenzhao.,Dong, Yinpeng.,Zhang, Xingxing.,Wang, Liyuan.,...&Su, Hang.(2025).RobustPrompt: Learning to defend against adversarial attacks with adaptive visual prompts.PATTERN RECOGNITION LETTERS,190,161-168.
MLA Liu, Chang,et al."RobustPrompt: Learning to defend against adversarial attacks with adaptive visual prompts".PATTERN RECOGNITION LETTERS 190(2025):161-168.
条目包含的文件
条目无相关文件。
个性服务
推荐该条目
保存到收藏夹
查看访问统计
导出为Endnote文件
谷歌学术
谷歌学术中相似的文章
[Liu, Chang]的文章
[Xiang, Wenzhao]的文章
[Dong, Yinpeng]的文章
百度学术
百度学术中相似的文章
[Liu, Chang]的文章
[Xiang, Wenzhao]的文章
[Dong, Yinpeng]的文章
必应学术
必应学术中相似的文章
[Liu, Chang]的文章
[Xiang, Wenzhao]的文章
[Dong, Yinpeng]的文章
相关权益政策
暂无数据
收藏/分享
所有评论 (0)
暂无评论
 

除非特别说明,本系统中所有内容都受版权保护,并保留所有权利。