
 
valuable only if the initial low-resolution images are 
blur-free and focused, stressing already the bad 
influence of low quality images in the fusion. In 
(Hollingsworth et al., 2009), authors proposed to 
perform a simple averaging of the normalized iris 
images extracted from the video for matching NIR 
videos against NIR videos from the MBGC 
database. When compared to a fusion of scores, the 
results are similar but with a reduced complexity. In 
the same spirit, Nguyen et al., (2010; 2011b) 
proposed to fuse different images of the video at a 
pixel level after an interpolation of the images. They 
use a quality factor in their fusion scheme, which 
allows giving less importance to images of bad 
quality. The interpolation step is shown very 
efficient as well as the quality weighting for 
improving recognition performance. Note that they 
considered a protocol similar to MBGC, where they 
compare a video to a high quality still image. More 
recent papers (Nguyen et al., 2011a); (Jillela et al., 
2011) explored the fusion in the feature domain 
using PCA or PCT but not on the same MBGC 
protocol as they usually degrade artificially the 
image resolution in their assessment stage. 
In our work, like in (Nguyen et al., 2011b), we 
propose to fuse the different frames of the video at 
the pixel level, after an interpolation stage which 
allows increasing the size of the resulting image by a 
factor of 2. Contrary to (Nguyen et al., 2011b), we 
do not follow the MBGC protocol which compares a 
video to a still high quality image reference but we 
consider in our work, a video against video scenario, 
more adapted to the re-identification context, 
meaning that we will use several frames in both low 
quality videos to address the person recognition 
hypothesis. 
The above literature review dealing with super 
resolution in the iris on the move context has 
stressed the importance of choosing adequately the 
images involved in the fusion process. Indeed, 
integration of low quality images leads to a decrease 
in performance producing a rather counterproductive 
effect. 
In this work, we will therefore concentrate our 
efforts in the proposition of a novel way of 
measuring and integrating quality measures in the 
image fusion scheme. More precisely our first 
contribution is the proposition of a global quality 
measure for normalized iris images as defined in 
(Cremer et al., 2012) as a weighting factor in the 
same way as proposed in (Nguyen et al., 2011b). 
The interest of our measure compared to (Nguyen et 
al., 2011b) is its simplicity and the fact that its 
computation does not require to identify in advance 
the type of degradations that can occur. Indeed our 
measure is based on a local GMM-based 
characterization of the iris texture. Bad quality 
normalized iris images are therefore images 
containing a large part of non-textured zones, 
resulting from segmentation errors or blur. 
Taking benefit of this local measure, we propose 
as a second novel contribution to perform a local 
weighting in the image fusion scheme, allowing this 
way to take into account the fact that degradations 
can be different in different parts of the iris image. 
This means that regions free from occlusions will 
contribute more in the reconstruction of the fused 
image than regions with artifacts such as eyelid or 
eyelash occlusion and specular reflection. Thus, the 
quality of the reconstructed image will be better and 
we expect this scheme to lead to a significant 
improvement in the recognition performance. 
This paper is organized as follows: in Section 2 
we describe our approach for Local and Global 
quality based super resolution and in Section 3 we 
present the comparative experiments that we 
performed on the MBGC database. Finally, 
conclusions are given in Section 4. 
2 LOCAL AND GLOBAL 
QUALITY-BASED SUPER 
RESOLUTION 
In this Section, we will first briefly describe the 
different modules of a video-based iris recognition 
system. We will also recall the definition of the local 
and global quality measure that we will use on the 
normalized images. This concept has been described 
in details in (Cremer et al., 2012); (Krichen et al., 
2007). We will explain how we have adapted this 
measure to the context of iris images resulting from 
low quality videos. We also describe the super-
resolution process allowing interpolation and fusion 
of the frames of the video. Finally, we will 
summarize the global architecture of the system that 
we propose for person recognition from moving 
person’s videos using these local and global quality 
measures. 
2.1  General Structure of Our 
Video-based Iris Verification 
System 
For building an iris recognition system starting from 
a video, several steps have to be performed. The first 
need is the detection and tracking of the eyes in the 
ICPRAM2013-InternationalConferenceonPatternRecognitionApplicationsandMethods
624