A Green AI Methodology Based on Persistent Homology for Compressing BERT
Metadatos
Mostrar el registro completo del ítemEditorial
MDPI
Materia
BERT compression Green AI persistent homology
Fecha
2025-01-03Referencia bibliográfica
Balderas Ruíz, L. & Lastra Leidinger, M. & Benítez Sánchez, J.M. Appl. Sci. 2025, 15, 390 [https://doi.org/10.3390/app15010390]
Patrocinador
Project with reference PID2020-118224RB-100 (funded by MICIU/AEI/10.13039/501100011033); Project PID2023-151336OB-I00 granted by the Spanish Ministerio de Ciencia, Innovación y UniversidadesResumen
Large Language Models (LLMs) like BERT have gained significant prominence
due to their remarkable performance in various natural language processing tasks. However,
they come with substantial computational and memory costs. Additionally, they are
essentially black-box models, being challenging to explain and interpret. In this article, Persistent
BERT Compression and Explainability (PBCE) is proposed, a Green AI methodology
to prune BERT models using persistent homology, aiming to measure the importance of
each neuron by studying the topological characteristics of their outputs. As a result, PBCE
can compress BERT significantly by reducing the number of parameters (47% of the original
parameters for BERT Base, 42% for BERT Large). The proposed methodology has been
evaluated on the standard GLUE Benchmark, comparing the results with state-of-the-art
techniques achieving outstanding results. Consequently, PBCE can simplify the BERT
model by providing explainability to its neurons and reducing the model’s size, making it
more suitable for deployment on resource-constrained devices.