
There  is  a  lot  of  room  to  improve  the  actual 
accuracy  of  the  system  -  we  might  be  able  to  use 
more  sophisticated  face  detection  algorithms  or 
classifiers, and even use techniques of hallucinating 
exemplars  from  the  existing  data,  to  make  the 
system  more  robust  to  noise  and  illumination 
conditions.  Nevertheless,  we  can  strongly  declare 
that our objective in this paper has been reached — 
it is  technically  possible to  make  a real-time  robust 
face  recognition  system  running  exclusively  on  the 
low-performance hardware of the smartwatch. 
Additionally,  in  terms  of  user  interaction,  the 
experiment  was  important  to  show  usability  and 
ergonomic  issues  that  need  to  be  addressed  before 
people with actual visual impairments are involved. 
The  feedback  that  indicates  a  face  is  being  framed 
needs more work so that it becomes a more precise 
clue as to where the user needs to point the 
smartwatch’s camera.  This  is  important  not  only  to 
allow  the  system  to  be  used  as  an  assistive 
technology,  but  also  to  alleviate  the  fatigue  issue 
reported  by  the  participants.  Other  potential  place 
for  future  enhancement  concerns  the  feedback 
interface to get data from people´s faces, which still 
must be  made  accessible for  use  by  blind  and low-
vision people. 
Finally,  we  propose  challenges  for  future  work, 
including  wearable  systems  for  objects  recognition, 
textual information recognition (e.g. signs, symbols) 
and a gesture recognition like Porzi et al. (2013), but 
processed within the smartwatch itself. Furthermore, 
we  will  conduct  experiments  to  better  analyze  the 
system's  energy  consumption.  Also,  experiments 
with  visually  impaired  users will  be  used  to  further 
evaluate  and  improve  the  system  as  an  assistive 
device. 
ACKNOWLEDGEMENTS 
The authors wish to express their gratitude to all the 
volunteers  who  participated  in  the  experiments  in 
this  study,  and  also  for  Samsung  Research  that 
loaned  the  hardware  equipment.  LSBN  receives  a 
Ph.D. fellowship from CNPq (grant  #141254/2014-
9).  VRMLM  receives  a  Ph.D.  fellowship  from 
CAPES  (grant  #01-P-04554/2013).  MCCB,  ARR 
and  SKG  receives  a  Productivity  Research 
Fellowship  from  CNPq  (grants  #308618/2014-9, 
#304352/2012-8 and #308882/2013-0, respectively). 
This work is part of a project that was approved by 
Unicamp  Institutional  Review  Board  CAAE 
31818014.0.0000.5404. 
REFERENCES 
Ahonen,  T.,  Hadid,  A.,  and  Pietikainen,  M.  (2006).  Face 
description  with  local  binary  patterns:  Application  to 
face  recognition.  IEEE Transactions on Pattern 
Analysis and Machine Intelligence,  28(12):2037–
2041. 
Astler, D., Chau, H., Hsu, K., Hua, A., Kannan, A., Lei, L. 
Nathanson,  M.,  Paryavi,  E.,  Rosen,  M.,  Unno,  H., 
Wang, C., Zaidi, K., Zhang, X., and  Tang, C. (2011). 
Increased  accessibility  to  nonverbal  communication 
through facial and expression recognition technologies 
for  blind/visually  impaired  subjects.  In  The 
Proceedings of the 13th International ACM 
SIGACCESS Conference on Computers and 
Accessibility, pages 259–260. 
Belhumeur,  P.,  Hespanha,  J.,  and  Kriegman,  D.  (1997). 
Eigenfaces  vs.  fisherfaces:  recognition  using  class 
specific  linear  projection.  IEEE Transactions on 
Pattern Analysis and Machine Intelligence, 
19(7):711–720. 
Chen,  X.,  Flynn,  P.,  and  Bowyer,  K.  (2003).  PCA-based 
face  recognition  in  infrared  imagery:  baseline  and 
comparative  studies.  In  Proceedings of the IEEE 
International Workshop on Analysis and Modeling of 
Faces and Gestures, pages 127–134. 
Cover,  T.,  and  Hart,  P.:  Nearest  neighbor  pattern 
classification.  (1967).  IEEE Transactions on 
Information Theory, 13(1):21–27. 
Dalal,  N.  and  Triggs,  B.  (2005).  Histograms  of  oriented 
gradients  for  human  detection.  In  IEEE Computer 
Society Conference on Computer Vision and Pattern 
Recognition, pages 886–893. 
FreevoxTouch (2014). The only smart watch in the world 
for  the  visually  impaired.  Available: 
http://myfreevox.com/en/. 
Fusco,  G.,  Noceti,  N.,  and  Odone,  F.  (2012).  Combining 
retrieval  and  classification  for  real-time  face 
recognition.  In  2013 10th IEEE International 
Conference on Advanced Video and Signal Based 
Surveillance, pages 276–281. 
Gordon, G. (1991). Face recognition based on depth maps 
and  surface  curvature.  In  SPIE1570, Geometric 
methods in Computer Vision, pages 234–247. 
Hadid,  A.,  Pietikainen,  M.,  and  Li,  S.  (2007).  Learning 
personal specific facial  dynamics for face  recognition 
from  videos.  In  Analysis and Modeling of Faces and 
Gestures: Lecture Notes in Computer Science 4778, 
pages 1–15. 
Kistler,  D.  and  Wightman,  F.  (1992).  A  model  of  head-
related  transfer  functions  based  on  principal 
components  analysis  and  minimum-phase 
reconstruction.  Journal of the Acoustical Society of 
America, 91(3):1637–1647. 
Kramer,  K.,  Hedin,  D.,  and  Rolkosky,  D.  (2010). 
Smartphone based face recognition tool for  the  blind. 
In 32nd Annual International Conference of the IEEE 
Engineering in Medicine and Biology Society,  pages 
4538–4541. 
Krishna,  S.,  Little,  G.,  Black,  J.,  and  Panchanathan,  S. 
AWearableFaceRecognitionSystemBuiltintoaSmartwatchandtheVisuallyImpairedUser
11