
 
units, and thus is likely not to play a useful role in 
the task. 
4.2  Pure Error-driven Learning 
Figure 4 presents the activation-based receptive field 
and projective field analysis for  the  lower left unit 
group in  layer V1  when  pure error-driven  learning 
was  used  for  training  (no  Hebbian  injected). 
Compared  to  the  previous  figure,  there  are  many 
units here that have not developed any useful feature 
representation,  and  project  equally  strongly  to  a 
large number of output units. In addition, quite a few 
units are never activated during processing. This is 
probably  why  generalization  performance  suffered 
when  Hebbian  learning  was  excluded  from  the 
learning mix (Equation 3; Figures 4).  
5  CONCLUSIONS 
In this article, we have studied the effect of mixing 
error  driven  and  Hebbian  learning  in  bidirectional 
hierarchical  networks  for  object  recognition.  Error 
driven  learning  alone  is  a  powerful  learning 
mechanism which  could  solve  the  task  at  hand  by 
learning  to  relate  individual  pixels  in  the  input 
patterns to desired perceptual categories. However, 
handwritten  letters  are  intrinsically  noisy  as  they 
contain  small  variations  due  to  different 
handwritings,  and  this  increases  the  risk  for 
overfitting—especially so in large networks. Hence, 
there  is  a  risk  that  error-driven  learning  might  not 
give  optimal  generalization  performance  for  these 
networks.  
We  run  systematic  training  and  generalization 
tests on a handwritten letter recognition task using 
pure  error-driven  learning  as  compared  to  using  a 
mixture  of  error-driven  and  Hebbian  learning.  The 
simulations indicate that mixing Hebbian and error-
driven learning can be quite successful in terms of 
improving  the  generalization  performance  of 
bidirectional  hierarchical  networks  in  cases  when 
there  is  much noise  in  the  input, and  an increased 
risk for overfitting. 
Additionally,  we  also  believe  that  Hebbian 
learning can be a good candidate for generic, local 
feature extraction for image processing and pattern 
recognition  tasks.  In  contrast  to  pre-wired  feature 
detectors,  for  example,  Gabor-filters,  Hebbian 
leaning provides a more flexible means for detecting 
the  underlying  statistical  structure  of  the  input 
patterns as it has no a priori constraints on the size or 
shape of these local features. 
REFERENCES 
Callaway,  E.  M.,  2004.  Feedforward,  feedback  and 
inhibitory connections in primate visual cortex. Neural 
Network, 17, 625-632. 
Fukushima,  K.,  1993.  Improved  generalization  ability 
using  constrained  neural  network  architectures, 
Proceedings of the International Joint Conference on 
Neural Networks, 2049-2054. 
Fukushima,  K.,  2008.  Recent  advances  in  the 
neocognitron. Neural Information Processing, Lecture 
Notes  In  Computer  Science,  1041-1050,  
Berlin/Heidelberg: Springer Verlag. 
Hebb,  D.  O.,  1949.    The  Organization  of  Behavior:  A 
Neuropsychological Theory. New York: Wiley. 
O'Reilly,  R.  C.  and  Munakata,  Y.,  2000.  Computational 
Explorations  in  Cognitive  Neuroscience: 
Understanding  the  Mind  by  Simulating  the  Brain. 
Cambridge, MA: MIT Press. 
McClelland, J., 2005. How far can you go with Hebbian 
learning,  and  when  does  it  lead  you  astray?  In 
Munakata, Y. and Johnson, M.H. (eds) Attention and 
Performance XXI: Processes of Change in Brain and 
Cognitive  Development.  Oxford:  Oxford  University 
Press. 
O'Reilly, R., 2001. Generalization in interactive networks: 
The  benefits  of  inhibitory  competition  and  Hebbian 
learning. Neural Computation, 13, 1199-1241.  
Wallis, G. and Rolls, E. T., 1997. Invariant face and object 
recognition  in  the  visual  system.  Progress  in 
Neurobiology. 51(2), 167-194. 
Aisa, B., Mingus, B. and O'Reilly, R., 2008. The emergent 
neural modeling system. Neural Networks, 21, 1146-
1152. 
BIDIRECTIONAL HIERARCHICAL NEURAL NETWORKS - Hebbian Learning Improves Generalization
111