Large Vision-Language Model Security: A Survey

Chong Zhang, Zheng Fang, Haochen Xue, Chong Zhang, Mingyu Jin, Wujiang Xu, Dong Shu, Shanchieh Yang, Zhenting Wang, Dongfang Liu*

*Corresponding author for this work

Research output: Contribution to journalConference articlepeer-review

1 Downloads (Pure)

Abstract

In the domain of Large Vision-Language Models (LVLMs), securing these models has emerged as a critical issue for both researchers and practitioners. In this paper, we highlight and analyze the security-related issues of LVLMs, with a special emphasis on the reliability challenges in practical deployments. We begin by reviewing recent studies on threats like jailbreak and backdoor attacks, alongside discussing the potential countermeasures implemented to mitigate these risks. Additionally, we touch on real-world application problems, such as hallucinations and privacy leakages, as well as the ethical and legal related researches around them. We also outline the shortcomings observed in current studies and discuss directions for future research, with the aim of promoting LVLMs towards a safer direction. A curated list of LVLMs-security-related resources is also available at https://github.com/MingyuJ666/LVLM-Safety.
Original languageEnglish
Article number1
Pages (from-to)3
Number of pages22
JournalCommunications in Computer and Information Science
Volume2315
Publication statusPublished - 27 Dec 2024

Fingerprint

Dive into the research topics of 'Large Vision-Language Model Security: A Survey'. Together they form a unique fingerprint.

Cite this