Mostrar el registro sencillo del ítem

dc.contributor.authorSavran Kiziltepe, Rukiye
dc.contributor.authorGan, John Q.
dc.contributor.authorEscobar Pérez, Juan José 
dc.date.accessioned2024-02-12T13:36:07Z
dc.date.available2024-02-12T13:36:07Z
dc.date.issued2024-02-01
dc.identifier.citationR. S. Kiziltepe, J. Q. Gan and J. J. Escobar, "Integration of Feature and Decision Fusion With Deep Learning Architectures for Video Classification," in IEEE Access, vol. 12, pp. 19432-19446, 2024, doi: 10.1109/ACCESS.2024.3360929es_ES
dc.identifier.urihttps://hdl.handle.net/10481/89102
dc.description.abstractInformation fusion is frequently employed to integrate diverse inputs, including sensory data, features, or decisions, in order to leverage the advantageous relationships among various features and classifiers. This paper presents a novel approach for video classification using deep learning architectures, including ConvLSTM and vision transformer based fusion architectures, which incorporates the combination of spatial and temporal features, along with the utilisation of decision fusion at multiple levels. The proposed vision transformer based method uses a 3D CNN to extract spatio-temporal information and different attention mechanisms to pay attention to essential features for action recognition and thus learns spatio-temporal dependencies effectively. The effectiveness of the methods proposed in this paper is validated through empirical evaluations conducted on two well-known video classification datasets, namely UCF-101 and KTH. The experimental findings indicate that the utilisation of both spatial and temporal features is essential, with the superior performance gained by using temporal features as the primary source of features in conjunction with two types of distinct spatial features when compared to other configurations. The multi-level decision fusion approach proposed in this study produces results comparable to those of feature fusion methods while offering the advantage of reduced memory requirements and computational costs. The fusion of RGB, HOG, and optical flow representations has demonstrated the best performance compared to other fusion methods examined in this study. It has also been demonstrated that the vision transformer based approaches significantly outperformed the ConvLSTM based approaches. Furthermore, an ablation study was conducted to compare the performances of vision transformer based feature fusion approaches for enhancing the performance of video classification.es_ES
dc.description.sponsorshipMinistry of National Education, Turkeyes_ES
dc.language.isoenges_ES
dc.publisherIEEEes_ES
dc.rightsCreative Commons Attribution-NonCommercial-NoDerivs 3.0 License
dc.rights.urihttp://creativecommons.org/licenses/by-nc-nd/3.0/
dc.subjectComputer visiones_ES
dc.subjectdata fusiones_ES
dc.subjectdeep neural networkses_ES
dc.subjecthuman action recognitiones_ES
dc.subjectspatio-temporal featureses_ES
dc.titleIntegration of Feature and Decision Fusion with Deep Learning Architectures for Video Classificationes_ES
dc.typejournal articlees_ES
dc.rights.accessRightsopen accesses_ES
dc.type.hasVersionVoRes_ES


Ficheros en el ítem

[PDF]

Este ítem aparece en la(s) siguiente(s) colección(ones)

Mostrar el registro sencillo del ítem

Creative Commons Attribution-NonCommercial-NoDerivs 3.0 License
Excepto si se señala otra cosa, la licencia del ítem se describe como Creative Commons Attribution-NonCommercial-NoDerivs 3.0 License