
and place’ task, serves as the primary evidence of the
system’s robustness. The 98% success rate (49 out
of 50 attempts) corroborates the effectiveness of the
processing chain, from motion capture to execution
by the actuator. Additionally, the low standard devi-
ation of 0.0126 seconds in cycle time attests to the
high precision and repeatability of the method, indis-
pensable characteristics for any industrial or research
application that demands consistency.
The implications of these results point to a notable
potential for application in real-world scenarios. In
the industrial sector, especially for small and medium-
sized enterprises, the technology can enable low-cost
automation for assembly, packaging, or quality con-
trol tasks, where flexibility and rapid reprogramming
are more critical than extreme speed. In the research
field, the system presents itself as a rapid prototyp-
ing platform for studies in Human-Robot Interaction
(HRI), allowing for the testing of new communication
and control modalities. Its educational value is also
considerable, serving as a practical tool for teaching
concepts in kinematics, automation, and computer vi-
sion in an applied and engaging manner.
In summary, the main contribution of this work
lies in the empirical validation of a solution that inte-
grates advanced computer vision to create an effective
and accessible human-robot interface. The research
demonstrates that it is feasible to abstract the com-
plexity of robotic programming, offering a control
method that is both powerful in its precision and sim-
ple in its use, representing a practical advancement in
the pursuit of more flexible and human-centered au-
tomation systems.
Future directions for this work focus on evolv-
ing the system beyond simple repetition, aiming for
greater intelligence and flexibility. The next step will
be the implementation of conditional operations, al-
lowing the robot to perform different actions based
on specific gestures. In parallel, the expansion of the
movement repertoire will be pursued to include more
complex tasks, such as contour following. A crucial
objective will also be the abstraction of the software
layer to ensure the solution’s portability, enabling its
adaptation to control different models of robotic arms
with minimal reconfiguration.
ACKNOWLEDGEMENTS
The project is supported by the National Council for
Scientific and Technological Development (CNPq)
under grant number 407984/2022-4; the Fund for
Scientific and Technological Development (FNDCT);
the Ministry of Science, Technology and Innovations
(MCTI) of Brazil; Brazilian Federal Agency for Sup-
port and Evaluation of Graduate Education (CAPES);
the Araucaria Foundation; the General Superinten-
dence of Science, Technology and Higher Education
(SETI); and NAPI Robotics.
REFERENCES
Amprimo, G., Masi, G., Pettiti, G., Olmo, G., Priano, L.,
and Ferraris, C. (2024). Hand tracking for clini-
cal applications: validation of the google mediapipe
hand (gmh) and the depth-enhanced gmh-d frame-
works. Biomedical Signal Processing and Control,
96:106508.
Bensaadallah, M., Ghoggali, N., Saidi, L., and Ghoggali,
W. (2023). Deep learning-based real-time hand land-
mark recognition with mediapipe for r12 robot con-
trol. In 2023 International Conference on Electri-
cal Engineering and Advanced Technology (ICEEAT),
volume 1, pages 1–6. IEEE.
Bradski, G. (2025). The OpenCV Library. Dr. Dobb’s Jour-
nal of Software Tools.
Du, J., Vann, W., Zhou, T., Ye, Y., and Zhu, Q. (2024). Sen-
sory manipulation as a countermeasure to robot tele-
operation delays: system and evidence. Scientific Re-
ports, 14(1):4333.
Franzluebbers, A. and Johnson, K. (2019). Remote robotic
arm teleoperation through virtual reality. In Sympo-
sium on Spatial User Interaction, pages 1–2.
Gomase, K., Dhanawade, A., Gurav, P., Lokare, S., and
Dange, J. (2022). Hand gesture identification using
mediapipe. International Research Journal of Engi-
neering and Technology (IRJET), 9(03).
Google (2025). Mediapipe: A framework for building per-
ception pipelines.
Javaid, M., Haleem, A., Singh, R. P., and Suman, R. (2021).
Substantial capabilities of robotics in enhancing in-
dustry 4.0 implementation. Cognitive Robotics, 1:58–
75.
John, T. S. (2011). Advancements in robotics and its future
uses. International Journal of Scientific & Engineer-
ing Research, 2(8):1–6.
Martinelli, D., Cerbaro, J., Fabro, J. A., de Oliveira, A. S.,
and Teixeira, M. A. S. (2020). Human-robot interface
for remote control via iot communication using deep
learning techniques for motion recognition. In 2020
Latin American Robotics Symposium (LARS), 2020
Brazilian Symposium on Robotics (SBR) and 2020
Workshop on Robotics in Education (WRE), pages 1–
6. IEEE.
Moniruzzaman, M., Rassau, A., Chai, D., and Islam,
S. M. S. (2022). Teleoperation methods and en-
hancement techniques for mobile robots: A compre-
hensive survey. Robotics and Autonomous Systems,
150:103973.
Moustris, G. P., Hiridis, S. C., Deliparaschos, K. M.,
and Konstantinidis, K. M. (2011). Evolution of au-
tonomous and semi-autonomous robotic surgical sys-
Real-Time Hand Gesture Control of a Robotic Arm with Programmable Motion Memory
299