
 
colored blob. In this step, we choose the biggest 
region as an object-colored blob. 
2.2  Adaptive Learning System 
Training is an off-line procedure that does not affect 
the on-line performance of the tracker. Nevertheless, 
the compilation of a sufficiently representative 
training set is a time-consuming and labor-intensive 
process. To cope with this problem, an adaptive 
training procedure has been developed. Training is 
performed on a small set of seed images for which a 
human provides ground truth by defining object-
colored regions. Following this, detection together 
with hysteresis thresholding is used to continuously 
update the prior probabilities P(o),  P(c) and P(c|o) 
based on a larger image data set. The updated prior 
probabilities are used to classify pixels of these 
images into object-colored and non-object-colored 
ones. The final training of the classifier is then 
performed based on the training set resulting from 
user editing. This process for adapting the prior 
probabilities  P(o),  P(c) and P(c|o) can either be 
disabled as soon as the achieved training is deemed 
sufficient for the purposes of the tracker, or continue 
as more input images are fed to the system. 
The success of the color detection depends 
crucially on whether or not the luminance conditions 
during the on-line operation of the detector are 
similar to those during the acquisition of the training 
data set. Despite the fact that using the UV color 
representation model has certain luminance 
independent characteristics, the object color detector 
may produce poor results if the luminance 
conditions during on-line operation are considerably 
different to those used in the training set. Thus, a 
means of adapting the representation of object-
colored image pixels according to the recent history 
of detected colored pixels is required. To solve this 
problem, object color detection maintains two sets of 
prior probabilities (Zabulis et al, 2009). The first set 
consists of P(o),  P(c),  P(c|o) that have been 
computed off-line from the training set. The second 
is made up of 
)(oP
W
, 
)(cP
W
, 
)|( ocP
W
, 
corresponding to the P(o),  P(c),  P(c|o) that the 
system gathers during the W most recent frames 
respectively. Obviously, the second set better 
reflects the “recent” appearance of object-colored 
objects and is therefore better adapted to the current 
luminance conditions. Object color detection is then 
performed based on the following moving average 
formula: 
(|) (|) ((1 ) (|))
AW
oc Poc P oc
γ
=+−
,
 
(2)
where 
)|( coP
A
 represents the adapted probability 
of a color c being an object color. P(o|c) and 
)|( coP
W
 are both given by Equation (1) but 
involve prior probabilities that have been computed 
from the whole training set [for P(o|c)] and from the 
detection results in the last W frames [for 
)|( coP
W
]. 
 is a sensitivity parameter that controls the 
influence of the training set in the detection process 
)10(
. If 
1
, then the object color detection 
takes into account only the training set (35 images in 
the off-line training set), and no adaptation takes 
place; if 
  is close to zero, then the object color 
detection becomes very reactive, relying strongly on 
the recent past for deriving a model of the immediate 
future. W is the number of history frames. If W value 
is too high, the length of history frames will be too 
long; if W value is set too low, the history for 
adaptation will be too short. In our implementation, 
we set 
= 0.8 and W = 5 which gave good results 
in the tests that have been carried out. 
Thus, the object color probability can be determined 
adaptively. By using on-line adaptation of object 
color probabilities, the classifier is easily able to 
cope with considerable luminance changes, and also 
it is able to segment the object even in the case of a 
dynamic background. 
3 RESULTS 
In this section, representative results from our 
experiment are shown. Figure 2 provides a few 
representative snapshots of the experiment. The 
reported experiment is based on a sequence that has 
been acquired with USB camera at a resolution of 
320x240 pixels. This process is done in real time 
and on-line. The experimental room is outdoor area 
(balcony). Note that the training set (35 images in 
the off-line training set) was collected from the 
indoor area. So the luminance change makes it much 
more challenging. 
The left window depicts the input images. The 
middle window shows the output images. The right 
window represents the object probability map in the 
U and V axis in color model, as depicted in Figure 3. 
In the initial stage (frame 15), when the 
experiment starts, the object color probability does 
not converge to a proper value. In other words, the 
color probability is scattering. So the segmented 
output cannot be achieved well because it uses only 
from   the  off-line  data  set  which  the  lighting  is 
SIMULTECH 2011 - 1st International Conference on Simulation and Modeling Methodologies, Technologies and
Applications 
202