From data to detection: Developing a corpus and training language models for the identification of anti-refugee narratives in Spanish
Metadatos
Mostrar el registro completo del ítemEditorial
Elsevier
Materia
Deep learning Language models Transformers
Fecha
2025-12Referencia bibliográfica
Mata, J., Gualda, E., Pachón, V., Rebollo, C., & Domínguez, J. L. (2025). From data to detection: Developing a corpus and training language models for the identification of anti-refugee narratives in Spanish. Array (New York, N.Y.), 28(100526), 100526. https://doi.org/10.1016/j.array.2025.100526
Patrocinador
MCIN/AEI/10.13039/501100011033 - FEDER/EU (PID2021-123983OB-I00, NON CONSPIRA HATE!); MCIN/AEI/10.13039/501100011033 , European Union – NextGenerationEU/PRTR (JDC2022-048239-I)Resumen
This study addresses the automatic detection of negative anti-refugee messages in Spanish texts, using language
models based on pre-trained Transformers models. Despite numerous studies on hate speech detection, few
have concentrated on Spanish, particularly regarding hostility towards refugees. To fill this void, we developed
HateRADAR-es, a new corpus of Spanish-language tweets manually annotated by sociologist and social workers
experts to identify the presence or absence of hateful content directed at refugees. This dataset has been
made available to the research community to encourage further investigation. A comprehensive experimental
framework to tackle this challenge, composed of several stages to achieve language models with a high
efficacy in detecting such messages, is presented. To address the class imbalance issue in the data, data
augmentation techniques are applied, and extensive experimentation is carried out to find the best values
for the hyperparameters of the language models to achieve better performance. In the evaluation process, an
ensemble of the fine-tuned models BETO, XLM-RoBERTa, and RoBERTa-large achieved the best results, with
an accuracy of 0.891, an F1-measure of 0.860, and an AUC-ROC of 0.892. These findings underscore the
effectiveness of combining multiple models into an ensemble to handle the complexity and nuances of hate
speech on social media, offering a promising direction for future adaptations and applications of language
models in specific hate contexts.