Abstract
Large Language Models (LLMs) are increasingly equipped with capabilities of real-time web search and integrated with protocols like Model Context Protocol (MCP). This extension could introduce new security vulnerabilities. We present a systematic investigation of LLM vulnerabilities to hidden adversarial prompts through malicious font injection in external resources like webpages, where attackers manipulate code-to-glyph mapping to inject deceptive content which are invisible to users. We evaluate two critical attack scenarios: (1) "malicious content relay" and (2) "sensitive data leakage" through MCP-enabled tools. Our experiments reveal that indirect prompts with injected malicious font can bypass LLM safety mechanisms through external resources, achieving varying success rates based on data sensitivity and prompt design. Our research underscores the urgent need for enhanced security measures in LLM deployments when processing external content.
| Original language | English |
|---|---|
| Title of host publication | The 2025 Conference on Empirical Methods in Natural Language Processing |
| Subtitle of host publication | EMNLP 2025 |
| Editors | Christos Christodoulopoulos, Tanmoy Chakraborty, Carolyn Rose, Violet Peng |
| Place of Publication | Suzhou, China |
| Publisher | Association for Computational Linguistics (ACL) |
| Chapter | 2025.findings-emnlp |
| Pages | 7133-7147 |
| Number of pages | 14 |
| Volume | 2505 |
| Edition | 16957 |
| Publication status | Published - 3 Nov 2025 |
| Event | The Conference on Empirical Methods in Natural Language Processing 2025: EMNLP - Suzhou International Expo Center, Suzhou, China, Suzhou, China Duration: 5 Nov 2025 → 9 Nov 2025 https://2025.emnlp.org/ |
Publication series
| Name | Association for Computational Linguistics |
|---|
Conference
| Conference | The Conference on Empirical Methods in Natural Language Processing 2025 |
|---|---|
| Country/Territory | China |
| City | Suzhou |
| Period | 5/11/25 → 9/11/25 |
| Internet address |