A Domain-Based Taxonomy of Jailbreak Vulnerabilities in Large Language Models
Metadatos
Mostrar el registro completo del ítemAutor
Peláez-González, Carlos; Herrera-Poyatos, Andrés; Zuheros, Cristina; Herrera-Poyatos, David; Tejedor, Virilo; Herrera Triguero, FranciscoEditorial
Elsevier
Materia
AI Safety Jailbreak LLMs
Fecha
2026-04-13Referencia bibliográfica
Published version: Peláez-González, C.; Herrera-Poyatos, A.; Zuheros, C. [et al]. (2026). A Domain-Based Taxonomy of Jailbreak Vulnerabilities in Large Language Models. Neurocomputing Volume 683, 133534. https://doi.org/10.1016/j.neucom.2026.133534
Patrocinador
National Institute of Cybersecurity and the University of Granada (IAFERCib C074/23); European Union (Next Generation)Resumen
The study of large language models (LLMs) is a key area in open-world machine learning. Although LLMs demonstrate remarkable natural language processing capabilities, they also face several challenges, including consistency issues, hallucinations, and jailbreak vulnerabilities. Jailbreaking refers to the crafting of prompts that bypass alignment safeguards, leading to unsafe outputs that compromise the integrity of LLMs. This work specifically focuses on the challenge of jailbreak vulnerabilities and introduces a novel taxonomy of jailbreak attacks grounded in the training domains of LLMs. It characterizes alignment failures as arising from gaps in generalization, objectives, and robustness.
Our primary contribution is a perspective on jailbreak, framed through the different linguistic domains that emerge during LLM training and alignment. This viewpoint highlights the limitations of existing approaches and enables us to classify jailbreak attacks in terms of the underlying model deficiencies they exploit.
Unlike conventional classifications that categorize attacks based on prompt construction methods (e.g., prompt templating), our approach provides a deeper understanding of LLM behavior. We introduce a taxonomy with four categories —mismatched generalization, competing objectives, adversarial robustness, and mixed attacks— offering insights into the fundamental nature of jailbreak vulnerabilities. Finally, we present key lessons derived from this taxonomic study.





