performed for each gesture independently as the en-
tire recognition flow. The final decision is achieved 
through a decision fusion. 
The remainder of this paper is organized as fol-
lows. Section 2 reviews template matching methods 
based on the longest common subsequence (LCSS). 
Section 3 describes the proposed method. Next, sec-
tion 4 exposes data sets used for our experiments. 
Section 5 reviews and discusses the results obtained. 
Finally, Section 6 draws a conclusion. 
2  RELATED WORK 
In this paper we present a new approach, based on the 
LM-WLCSS. Our state of the art briefly summarizes 
related methods. We first introduce the training phase 
of the Segmented and Warping LCSS. Then, the 
recognition phase of these methods are described re-
spectively. We also present the LM-WLCSS as our 
technique is derived from it. Finally, we present our 
contributions to the LM-WLCSS. 
Templates matching methods (TMMs) (Hart-
mann and Link, 2010) based on Dynamic Time Warp-
ing (Hartmann and Link, 2010), were demonstrated 
as non-efficient in presence of noisy raw signals (Vla-
chos et al., 2003). To handle such data, Long-Van et 
al. (2012) have introduced two new methods, based 
on Longest Common Subsequence (LCSS), Segment-
edLCSS and WarpingLCSS. Both methods share the 
same training phase. This training allows converting 
accelerometer data into strings. This is due to the fact 
that LCSS is based on a problem that relies upon 
strings. For this, raw signals must be quantized. The 
quantization step, proposed in (Long-Van et al., 
2012), involves computing clusters upon the training 
data with the K-Means algorithm. Those centroids are 
associated with pre-defined symbols to form strings. 
Therefore, each gesture instance is represented as a 
sequence of symbols. A LCSS score is associated 
with each sequence. The higher the LCSS score is be-
tween two elements, the greater the similarity is. A 
gesture instance is thus defined as a temporary tem-
plate. The final motif is chosen based on the one with 
the highest average LCSS score. However, in order to 
be able to compute whether a signal belongs to a ges-
ture class or not, a rejection threshold is associated 
with the template. This threshold is defined as the 
minimum LCSS between the previously elected tem-
plate and all other gesture instances of the same class. 
Yet, Nguyen-Dinh et al. (2014b) have suggested a 
new rejection threshold calculation, based on the 
mean μ
c
 and standard deviation σ
c
 of LCSS scores for 
the given class c. The resulting threshold ε is defined 
as ε= μ
c
-h·σ
c
, where h is an integer that allows ad-
justing the sensitivity of the algorithm for this class. 
In the Segmented LCSS recognition process, the 
stream is stored in a sliding window OW. Each sam-
ple of this window is associated with previously gen-
erated centroids and its related symbol, based on the 
minimum Euclidean distance. Then, this new string is 
entirely compared to the template computed during 
the training. If the resulting score exceeds the rejec-
tion threshold, of the associated class, then the gesture 
is associated to c. However, a gesture may be spotted 
as belonging to more than one class. To resolve such 
conflicts, a resolver may be added, as proposed in 
(Long-Van et al., 2012). It is based on the normalized 
similarity 
NormSim(A, B) = LCSS(A,B)/max(
‖
A
‖
,
‖
B
‖
), where 
‖
‖
 and 
‖
‖
 are respectively the length of A and B 
strings. The class with the highest NormSim is then 
marked as recognized. However, the Segmented 
LCSS method implies to recompute the score each 
time the sliding window is shifted. As a result, the 
computation time is O(T
2
) (with T the size of the 
longest template) in the worst case. However, without 
OW the LCSS algorithm cannot find boundaries of in-
coming gestures. In this way, Long-Van et al. (2012) 
have introduced a new variant of the LCSS called 
WLCSS (WLCSS). 
The WLCSS method removes need of a sliding 
window and improves the computational cost as it au-
tomatically determines gesture boundaries. In this 
new variant, quantized signals are still compared to 
the template of a given class. Nevertheless, this ver-
sion only update the score for each new element, 
starting from zero. This score grows when a match 
occurs and decreases thanks to penalties otherwise. 
The penalty consists of a weighted Euclidean distance 
between symbols, whether it is a mismatch, a repeti-
tion in the stream or even in the template. In a newer 
version presented in (Nguyen-Dinh et al., 2014b), the 
distance is normalized. Once the matching score is 
updated, the final result is output by the same decision 
maker used in the SegmentedLCSS method. The re-
sulting time complexity for this new method is O(T). 
Although the computational cost WLCSS is one order 
of magnitude lower than the SegmentedLCSS, the 
memory usage remains O(T
2
) in the worst case. 
Recently, Roggen et al. (2015) have proposed a 
new, microcontroller optimized, version of the 
WLCSS algorithm called Limited Memory and 
WLCSS (LM-WLCSS). Identically to previous meth-
ods, this one is designed to spot motif in noisy raw 
signals and focuses on a single sensor channel. In this 
way, a quantization step may not be required. More-
over, the training phase of this new variant has also