Paper ID: 2505.06913 • Published May 11, 2025
RedTeamLLM: an Agentic AI framework for offensive security
Brian Challita, Pierre Parrend
Laboratoire de Recherche de l’EPITA•ICube, UMR 7357, Universit´e de Strasbourg, CNRS
TL;DR
Get AI-generated summaries with premium
Get AI-generated summaries with premium
From automated intrusion testing to discovery of zero-day attacks before
software launch, agentic AI calls for great promises in security engineering.
This strong capability is bound with a similar threat: the security and
research community must build up its models before the approach is leveraged by
malicious actors for cybercrime. We therefore propose and evaluate RedTeamLLM,
an integrated architecture with a comprehensive security model for
automatization of pentest tasks. RedTeamLLM follows three key steps:
summarizing, reasoning and act, which embed its operational capacity. This
novel framework addresses four open challenges: plan correction, memory
management, context window constraint, and generality vs. specialization.
Evaluation is performed through the automated resolution of a range of
entry-level, but not trivial, CTF challenges. The contribution of the reasoning
capability of our agentic AI framework is specifically evaluated.
Figures & Tables
Unlock access to paper figures and tables to enhance your research experience.