Contextual Attack
Contextual attacks exploit the inherent context-awareness of large language models (LLMs) and other deep learning systems to bypass security measures and elicit undesired outputs. Current research focuses on developing sophisticated attack strategies that leverage subtle contextual cues, including implicit references, uncommon text structures, and multi-round interactions, to manipulate model behavior without triggering detection mechanisms. These attacks highlight vulnerabilities in existing defense mechanisms and underscore the need for more robust security protocols in AI systems, impacting the reliability and safety of applications across various domains, including IoT and aerial detection.
Papers
October 4, 2024
June 13, 2024
February 14, 2024
February 27, 2023
February 15, 2023
September 28, 2022
March 29, 2022