To assess the effectiveness of these approaches,
they were qualitatively evaluated in experiments on
a synthetic data set and two real-world data sets,
namely wine quality and KDD-Cup99 HTTP. The ex-
periments showed that all three approaches are suit-
able for increasing the explainability of the outlier de-
tection results by identifying the features which are
most relevant for the algorithm to detect the outliers.
Furthermore, it was found that the feature ranking re-
sults depend on the algorithm used. The GMM fo-
cuses strongly on linear relationships between the fea-
tures and is particularly suitable when the data can be
modeled by a fixed number of Gaussian components.
If this is not the case (e.g. the underlying distribu-
tion is not a Gaussian distribution), the GMM neglects
the relationship of different features to each other and
tends to explain global outliers only. This leads to a
feature ranking assuming independent features, which
is often not the case. The AE approach can model
by its non-linearity also various feature relationships.
Likewise, the k-NN approach is not bound to linear
relationships as well. This leads to a different feature
ranking that is more helpful in general, especially if
the underlying distribution is unknown.
Overall, all three approaches supports the task of
outlier analysis to better understand the results of the
algorithms and explain the outliers. Since many other
commonly used outlier detection algorithms are also
distance- or probability-based, this work can serve as
a basis for investigating further into the topic of ex-
plainable outlier detection using feature ranking.
REFERENCES
Aggarwal, C. C. (2017). Outlier analysis. Springer, second
edition.
Amer, M. and Goldstein, M. (2012). Nearest-neighbor
and clustering based anomaly detection algorithms for
rapidminer. In Proceedings of the 3rd RapidMiner
Community Meeting and Conferernce, pages 1–12.
An, J. and Cho, S. (2015). Variational autoencoder based
anomaly detection using reconstruction probability.
Special Lecture on IE, 2(1):1–18.
Angiulli, F. and Pizzuti, C. (2002). Fast Outlier Detection in
High Dimensional Spaces. In Principles of Data Min-
ing and Knowledge Discovery, volume 2431, pages
15–27. Springer Berlin Heidelberg.
Anscombe, F. J. (1960). Rejection of Outliers. Technomet-
rics, 2(2):123–146.
Antwarg, L., Miller, R. M., Shapira, B., and Rokach, L.
(2019). Explaining anomalies detected by autoen-
coders using shap.
Baxter, R. A. (2017). Mixture Model. In Encyclopedia of
Machine Learning and Data Mining, pages 841–844.
Springer US.
Burkart, N. and Huber, M. F. (2020). A survey on the ex-
plainability of supervised machine learning. CoRR,
abs/2011.07876.
Chandola, V., Banerjee, A., and Kumar, V. (2009).
Anomaly detection: A survey. ACM Computing Sur-
veys, 41(3):1–58.
Cortez, P., Cerdeira, A., Almeida, F., Matos, T., and Reis,
J. (2009). Modeling wine preferences by data mining
from physicochemical properties. Decision Support
Systems, 47(4):547–553.
Dempster, A. P., Laird, N. M., and Rubin, D. B. (1977).
Maximum Likelihood from Incomplete Data via the
EM Algorithm. Journal of the Royal Statistical Soci-
ety. Series B (Methodological), 39(1):1–38.
Goldstein, M. (2015). Unsupervised Anomaly Detection
Benchmark. https://doi.org/10.7910/DVN/OPQMVF.
Goldstein, M. and Uchida, S. (2016). A Comparative Evalu-
ation of Unsupervised Anomaly Detection Algorithms
for Multivariate Data. PLOS ONE, 11(4):e0152173.
Goodfellow, I., Bengio, Y., and Courville, A. (2016). Deep
Learning. MIT Press.
Grubbs, F. E. (1969). Procedures for Detecting Outlying
Observations in Samples. Technometrics, 11(1):1–21.
Hawkins, S., He, H., Williams, G. J., and Baxter, R. A.
(2002). Outlier Detection Using Replicator Neu-
ral Networks. In Proceedings of the 4th Interna-
tional Conference on Data Warehousing and Knowl-
edge Discovery, pages 170–180. Springer-Verlag.
Herskind Sejr, J., Christiansen, T., Dvinge, N., Hougesen,
D., Schneider-Kamp, P., and Zimek, A. (2021). Out-
lier detection with explanations on music streaming
data: A case study with danmark music group ltd. Ap-
plied Sciences, 11(5).
Lindsay, B. G. (1995). Mixture Models: Theory, Geometry
and Applications. NSF-CBMS Regional Conference
Series in Probability and Statistics, 5:i–163.
McLachlan, G. and Peel, D. (2000). Finite Mixture Models:
McLachlan/Finite Mixture Models. Wiley Series in
Probability and Statistics. John Wiley & Sons, Inc.
Panjei, E., Gruenwald, L., Leal, E., Nguyen, C., and Sil-
via, S. (2022). A survey on outlier explanations. The
VLDB Journal, 31(5):977–1008.
Ramaswamy, S., Rastogi, R., and Shim, K. (2000). Effi-
cient algorithms for mining outliers from large data
sets. ACM SIGMOD Record, 29(2):427–438.
Sejr, J. H. and Schneider-Kamp, A. (2021). Explainable
outlier detection: What, for whom and why? Machine
Learning with Applications, 6:100172.
Tavallaee, M., Bagheri, E., Lu, W., and Ghorbani, A. A.
(2009). A detailed analysis of the kdd cup 99 data set.
In 2009 IEEE Symp. on Computational Intelligence
for Security and Defense Applications, pages 1–6.
Explainable Outlier Detection Using Feature Ranking for k-Nearest Neighbors, Gaussian Mixture Model and Autoencoders
253