Abstract
t. This paper presents a sensor-based approach to developing a gesture-controlled robotic vehicle, designed
to enhance human-machine interaction in a more intuitive and accessible manner. The system translates hand
movements, detected through accelerometers and gyroscopes within a wearable device, into commands that control
the vehicle. Traditional methods for operating robotic systems, such as joysticks or remote controls, often require
physical interaction, limiting their effectiveness, especially for individuals with disabilities or in environments
where hands-free control is essential. This approach offers a non-contact, real-time control mechanism, detecting
hand gestures without relying on visual cues, thereby reducing computational load and ensuring stable performance
even under varying lighting conditions. The implementation involves sensor calibration, data acquisition, and
processing via microcontrollers such as Arduino or ESP32, coupled with a motor control system for vehicle
operation. Testing results demonstrate the accuracy and low-latency response of the gesture-based control system.
The primary objective is to make this technology widely accessible, offering a scalable solution for various
applications, including assistive technologies and robotics. Preliminary findings confirm the system’s effectiveness,
opening the door for further real-world applications.