
 
If 
t
tt
TxyP ,,≤
κ
, set 
tt
yx =
+1
 and 
tt
yfxf =
+1
; otherwise, set 
tt
xx =
+1
 and 
tt
xfxf =
+1
. 
Step 4.
 If the prescribed termination condition is 
satisfied, then stop; otherwise, update the value of 
the temperature by means of the temperature 
updating function, and then go back to Step 2. 
Thus, by applying the generation mechanism and 
the Metropolis acceptance criterion, the SA 
algorithm produces two sequences of random points. 
These are the sequence 
0, ≥ty
t
 of trial points 
generated by (4) and the sequence 
0, ≥tx
t
 of 
iteration points determined by applying the 
Metropolis acceptance criterion as described in Step 
3. These two sequences of random variables are all 
dependent on the temperature sequence 
}
0, ≥tT
t
 
determined by the temperature updating function, 
the state neighbouring sequence
{}
0, ≥t
t
, and the 
approach of random vector generation. 
The sequence 
{}
0, ≥t
t
 of positive numbers 
specified in Step 1 of the above SA algorithm is 
used to impose a lower bound on the random vector, 
generated at the each iteration, for obtaining the 
random trial point. This lower bound should be 
small enough and monotonically decreasing as the 
annealing proceeds. Since the temperature-
dependent generation probability density function is 
used to generate random trial points and since only 
one trial point is generated at each temperature value 
the SA algorithm considered is characterized by a 
nonhomogeneous continuous-state Markov chain.  
The convergence conditions of the SA were 
studied by Yang (Yang, 2000) and several updating 
functions for the method parameters were given, 
which ensure convergence of the method. We 
applied the next updating functions in testing our 
approach. 
Let 
n
r ℜ∈
, with component 
ii
Dyx
i
yxr −=
∈,
max
, 
ni ≤≤1
, 
1>d
, 
1>u
, 
u
0
, 
i
ni
r
≤≤
<<
1
0
min0
, 
nu
t
t
⋅
−
⋅=
λ
ρρ
0
 for all 
1≥t
, where 
{}
0, ≥t
t
 is the sequence used to impose lower 
bounds on the random vectors generated in the SA 
algorithm. Let the temperature-dependent generation 
probability density function 
()
t
Tp ,⋅
 be given by 
.,1log1
2
)1(
),(
1
n
d
t
i
n
i
t
i
t
t
z
T
z
T
z
T
a
Tzp ℜ∈
⎟
⎟
⎠
⎞
⎜
⎜
⎝
⎛
⎟
⎟
⎠
⎞
⎜
⎜
⎝
⎛
+
⎟
⎟
⎠
⎞
⎜
⎜
⎝
⎛
+⋅
−
=
∏
=
 
Then, for any initial point
Dx ∈
0
, the sequence 
0);( ≥txf
t
 of objective function values converges 
in probability to the global minimum
*
f
, if the 
temperature sequence 
}
0, ≥tT
t
 determined by the 
temperature updating function satisfies the following 
condition: 
⎟
⎟
⎟
⎠
⎞
⎜
⎜
⎜
⎝
⎛
⋅−⋅=
⋅nd
t
tlTT
1
0
exp
, 
...,,,i 21=
 
where 
0
0
>T
 is the initial temperature value and 
0>l
 is a given real number (Yang, 2000). 
Typically a different form of the temperature 
updating function has to be used with respect to a 
different kind of the generation probability density 
function in order to ensure the global convergence of 
the corresponding SA algorithm. Furthermore, the 
flatter is the tail of the generation probability 
function, the faster is the decrement of the 
temperature sequence determined by the temperature 
updating function. 
4 SVM CLASSIFICATION  
Data classification is a common problem in science 
and engineering. Support Vector Machines (SVMs) 
are powerful tools for classifying data that are often 
used in data mining operations.  
In the standard binary classification problem, a 
set of training data 
ii
y,u , … ,
mm
y,u  is 
observed, where the input set of points is 
ni
Uu ℜ⊂∈ , the 
i
y  is either +1 or −1, indicating 
the class to which the point 
i
u  belongs, 
}
11 −+∈ ,y
i
. The learning task is to create the 
classification rule 
{}
11 −+→ ,U:f  that will be 
used to predict the labels for new inputs. The basic 
idea of SVMs classification is to find a maximal 
margin separating hyperplane between two classes. 
It was first described by Cortes and Vapnik (Cortes 
& Vapnik, 1995). The standard binary SVM 
classification problem is shown visually in Figure 1. 
ICORES 2012 - 1st International Conference on Operations Research and Enterprise Systems
152