@misc{10481/103721, year = {2024}, month = {12}, url = {https://hdl.handle.net/10481/103721}, abstract = {Before the advent of machine learning and AI, systems predicting human intent and movement relied heavily on sensor-based approaches like inertial measurement units (IMUs), gyroscopes, and accelerometers, which primarily tracked physical movements. These systems, while effective in detecting motion, lacked the nuanced understanding of human intent and environmental context that could be gained from integrating human gaze. The title "Synergizing Human Gaze with Machine Vision for Location Mode Prediction" reflects the integration of human gaze data, which provides information about where a person is looking (indicating intent), with machine vision systems that process movement data (cloud points) to predict future locomotion modes or transitions. Before machine learning, traditional systems for predicting human movement were limited to sensor-based methods such as IMUs, which could only detect physical movements without understanding the intent behind them. These systems were less adaptable and often required manual calibration and interpretation by experts. Traditional sensor-based systems lacked the ability to accurately predict human intent or understand the contextual environment in real-time, leading to less reliable and slower responses in applications like wearable robotics. These systems could detect movement but were unable to forecast the user's next movement or transition. The proposed system, GT-NET, utilizes machine learning algorithms to combine human gaze data (images) with cloud point data (user movement) for predicting human intent and locomotion. This system leverages deep learning models trained on a custom dataset, with the aim of accurately forecasting the user's next movement. By integrating these data modalities, GT-NET enhances the ability of machines to anticipate human actions, particularly in dynamic environments.}, publisher = {Universidad de Granada}, keywords = {Machine learning}, keywords = {Artificial Intelligence (AI)}, keywords = {Deep Learning Models}, keywords = {Cloud}, title = {Synergizing Human Gaze with Machine Vision for Location Mode Prediction}, author = {Vanitha, M. and Niharika, N. and Pranitha, L. and Architha, K.}, }