AI Security Beyond Core Domains: Resume Screening as a Case Study of Adversarial Vulnerabilities in Specialized LLM Applications
By: Honglin Mu , Jinghao Liu , Kaiyang Wan and more
Potential Business Impact:
Makes AI safer from hidden trick instructions.
Large Language Models (LLMs) excel at text comprehension and generation, making them ideal for automated tasks like code review and content moderation. However, our research identifies a vulnerability: LLMs can be manipulated by "adversarial instructions" hidden in input data, such as resumes or code, causing them to deviate from their intended task. Notably, while defenses may exist for mature domains such as code review, they are often absent in other common applications such as resume screening and peer review. This paper introduces a benchmark to assess this vulnerability in resume screening, revealing attack success rates exceeding 80% for certain attack types. We evaluate two defense mechanisms: prompt-based defenses achieve 10.1% attack reduction with 12.5% false rejection increase, while our proposed FIDS (Foreign Instruction Detection through Separation) using LoRA adaptation achieves 15.4% attack reduction with 10.4% false rejection increase. The combined approach provides 26.3% attack reduction, demonstrating that training-time defenses outperform inference-time mitigations in both security and utility preservation.
Similar Papers
AI Hiring with LLMs: A Context-Aware and Explainable Multi-Agent Framework for Resume Screening
Computation and Language
Helps hire people faster by reading resumes.
Real-VulLLM: An LLM Based Assessment Framework in the Wild
Cryptography and Security
Finds computer bugs to make software safer.
When Reject Turns into Accept: Quantifying the Vulnerability of LLM-Based Scientific Reviewers to Indirect Prompt Injection
Artificial Intelligence
Tricks AI judges to accept bad science papers.