proposed system for applications such as animal
monitoring and agricultural surveillance. The model
was trained on a dataset of 50,000 images for 20
different animal species and managed to obtain 92%
accuracy and 89% adaptability rate in different
environmental conditions. The rationale behind these
findings is consistent with existing literature that
highlights the importance of diverse datasets for
enabling deep learning models in generalizing across
environments (Wang et al., 2021). One of the major
disadvantages of the system is that it is less accurate
with heavily occluded objects or unfamiliar species.
Its accuracy decreases to 73% for half covered
objects and 70% for species not seen during training.
Duplicate challenges have been observed in
predictive eye-tracking models, whereby data voids
and occlusions hamper the effectiveness of tracking
(Jyotsna et al., 2023).
As a part of this research, in order to increase
robustness of eye-ball tracking for human-computer
interaction, the data set can be extended with
additional 20,000 images of different movements.
Related to above mentioned eye-ball tracking, by
using existing images tracking can also be improved
by various methods. This can allow the model to run
on edge devices such as Raspberry Pi or NVIDIA
Jetson devices by reducing computational load by
30% and keeping over 90% accuracy of the model.
Deep learning-based approaches for gaze analysis
have recently been optimized to compress the model
size by up to 40% while keeping their speed in real-
time (Ghosh et al., 2024). Retinal fundus images with
multiple cameras enable the development of novel
eye-ball tracking approaches; however, the accuracy
of the proposed eye-ball tracking system is 30%
lower than the eye-ball tracking accuracy with a
single camera and is limited due to high
computational complexities and potentially achieving
a biased dataset under heavy occlusions and rapid
movements (Zuo et al., 2024). In future work, the
CNN model can be optimized to reduce
computational loads by 30% and 20,000 additional
images should be used to expand the dataset further
enhancing robustness.
7 CONCLUSIONS
An eye-ball tracking system based on CNN was built
and observed with 1,000 to 10,000 images. Training
was done in times of distinct real environmental
conditions like head orientation, occlusions, eye
conditions, and lighting levels. However, unlike the
other studies, in this study the CNN and RNN
comparison showed a better accuracy performance
for CNN of 96.5% with faster processing time thanks
to an optimized technique while the maximum
accuracy only for RNN was 80% with a slower
computation time. With processing times decreasing
with the growth of dataset size, the CNN become
capable of real-time operation and great stability as
well as efficiency for eye tracking applications in
practice.
REFERENCES
H. O. Edughele, Y. Zhang, F. Muhammad-Sukki, Q. -T.
Vien, H. Morris-Cafiero and M. Opoku Agyeman,
"Eye-Tracking Assistive Technologies for Individuals
With Amyotrophic Lateral Sclerosis," in IEEE Access,
vol. 10, pp. 41952 41972, 2022, doi:10.1109/ACCESS
.2022.3164075.
K. Iddrisu, W. Shariff, P. Corcoran, N. E. O'Connor, J.
Lemley and S. Little, "Event Camera-Based Eye
Motion Analysis: A Survey," in IEEE Access, vol. 12,
pp. 136783 136804, 2024, doi:10.1109/ACCESS.2024
.3462109.
R. Kannan Megalingam, S. Kuttankulangara Manoharan,
G. Riju and S. Makkal Mohandas, "Netravaad:
Interactive Eye Based Communication System for
People With Speech Issues," in IEEE Access, vol. 12,
pp. 69838 69852, 2024, doi:10.1109/ACCESS.2024.3
402334.
F. Zuo, P. Jing, J. Sun, J. Duan, Y. Ji and Y. Liu, "Deep
Learning-Based Eye-Tracking Analysis for Diagnosis
of Alzheimer's Disease Using 3D Comprehensive
Visual Stimuli," in IEEE Journal of Biomedical and
Health Informatics, vol. 28, no. 5, pp. 2781-2793, May
2024, doi: 10.1109/JBHI.2024.3365172.
Y. Wang, S. Lu and D. Harter, "Towards Collaborative and
Intelligent Learning Environments Based on Eye
Tracking Data and Learning Analytics: A Survey," in
IEEE Access, vol. 9, pp. 137991-138002, 2021, doi:
10.1109/ACCESS.2021.3117780.
C. Jyotsna, J. Amudha, A. Ram, D. Fruet and G. Nollo,
"PredictEYE: Personalized Time Series Model for
Mental State Prediction Using Eye Tracking," in IEEE
Access, vol. 11, pp. 128383-128409, 2023, doi:
10.1109/ACCESS.2023.3332762.
S. Ghosh, A. Dhall, M. Hayat, J. Knibbe and Q. Ji,
"Automatic Gaze Analysis: A Survey of Deep Learning
Based Approaches," in IEEE Transactions on Pattern
Analysis and Machine Intelligence, vol. 46, no. 1, pp.
61-84, Jan. 2024, doi: 10.1109/TPAMI.2023.3321337.
J. Santhosh, D. Dzsotjan and S. Ishimaru, "Multimodal
Assessment of Interest Levels in Reading: Integrating
Eye-Tracking and Physiological Sensing," in IEEE
Access, vol. 11, pp. 93994-94008, 2023, doi:
10.1109/ACCESS.2023.3311268.
R. Rathnayake et al., "Current Trends in Human Pupil
Localization: A Review," in IEEE Access, vol. 11, pp.