rare or if models are only moderately well defended. 
However, the impact of taste sampling is non-linear 
especially in  systems with  highly defended models. 
In  such  situations  taste-sampling  lowers
r *
. 
Consequently, under the assumption of a fixed value 
for 
r *
  and  stability,  a  predator  evolves  a  taste-
sampling strategy because mimics are less  common 
or models are better defended than in a comparable 
stable  environment  where  predators  do  not  utilize 
taste sampling.  
Another  interesting  aspect  is  the  effect  of 
different  age  distributions:  in  general  longevity  in 
predators  increases 
r *
.  The  effects  are  linear  with 
regards  to  prey  abundance  but  non-linear  with 
regards  to  prey  toxicity  where  behavioural 
expenditure  gains  increasing  impact  in  the  case  of 
defended  prey  and  older  predators,  whereas 
metabolic costs have an increased impact in the case 
of non-defended prey. The main conclusions of this 
paper are as follows: 
 
  On  the  predator’s  side 
r *
is  related  to  the 
nutritional value of prey and on the prey’s side 
it  relates  to  an  energy  inventory  which  can  be 
allocated,  amongst  other  things,  towards  the 
cost of defences or reproduction; 
  Behavioural  expenditure  has  a  greater  impact 
than metabolic costs when prey is rare and 
undefended; 
  Metabolic  costs  have  a  greater  impact  when 
prey is abundant or highly defended; 
  Longevity  of  the  predator  increases  the 
importance  of  behavioural  expenditure  in  the 
case of highly defended prey and the impact of 
metabolic costs if prey is undefended; 
  Mimics generally lower 
r *
 
which leads to less 
nutritional  prey  or  better  defended  models  if 
r *
is meant to be unchanged; 
  Predators  utilize  taste  sampling  if  mimics  are 
rare or models are highly toxic. 
REFERENCES 
Alonso,  E.,  Fairbank,  M.,  and  Mondragón,  E.  (2015). 
Back to optimality: A formal framework to express the 
dynamics  of  learning  optimal  behavior.  Adaptive 
Behavior, 23(4), 206-215. 
Barto,  A.G.,  Sutton,  R.S.,  and  Watkins,  C.J.C.H.  (1990). 
Learning and sequential decision making. In Learning 
and Computational Neuroscience: Foundations of 
Adaptive Networks, M. Gabriel and J.W. Moore, Eds., 
pp. 539-602, Cambridge, Mass: MIT Press.  
Dayan,  P.,  and  Daw,  N.D.  (2008).  Decision  theory, 
reinforcement  learning,  and  the  brain,  Cognitive, 
Affective, & Behavioral Neuroscience 8, 429–453. 
Dingemanse, N. J., and Réale, D. (2005). Natural selection 
and animal personality, Behavior 142, 1159–1184.  
Orr,  H.  A.,  (2009).  Fitness  and  its  role  in  evolutionary 
genetics. Nature Review Genetics 10, 531-539. 
Rangel,  A.,  Camerer,  C.,  and  Montague,  P.R.  (2008).  A 
framework  for  studying  the  neurobiology  of  value-
based decision making, Nature Reviews Neuroscience 
9, 545–556. 
Schultz, W. (2008). Neuroeconomics: the promise and the 
profit, Philosophical Transactions of the Royal Society 
B: Biological Sciences 363, 3767–3769.  
Staddon,  J.E.  (2007).  Is  animal  behavior  optimal?  In  A. 
Bejan  &  G.W.  Merkx  (eds.)  Constructal Theory of 
Social Dynamics, NY: Springer. 
Sutton,  R.S.,  and  Barto,  A.G.  (1998).  Reinforcement 
learning: An introduction, Boston, MA: Cambridge 
University Press.  
Teichmann, J. (2014). Models of aposematism and the role 
of aversive learning. PhD dissertation, City University 
London, London, UK. 
Teichmann,  J.,  Broom,  M.,  and  Alonso,  E.  (2014).  The 
application of temporal difference  learning in optimal 
diet  models, Journal of Theoretical Biology  340,  11–
16.