Towards Gaussian Processes
for Automatic and Interpretable Anomaly Detection in Industry 4.0
Fabian Berns
1
, Markus Lange-Hegermann
2
and Christian Beecks
1
1
Department of Computer Science, University of M
¨
unster, Germany
2
Department of Electrical Engineering and Computer Science,
OWL University of Applied Sciences and Arts, Lemgo, Germany
markus.lange-hegermann@th-owl.de
Keywords:
Anomaly Detection, Gaussian Processes, Explainable Machine Learning, Industry 4.0.
Abstract:
Discerning unexpected from expected data patterns is the key challenge of anomaly detection. Although a
multitude of solutions has been applied to this modern Industry 4.0 problem, it remains an open research issue
to identify the key characteristics subjacent to an anomaly, sc. generate hypothesis as to why they appear. In
recent years, machine learning models have been regarded as universal solution for a wide range of problems.
While most of them suffer from non-self-explanatory representations, Gaussian Processes (GPs) deliver inter-
pretable and robust statistical data models, which are able to cope with unreliable, noisy, or partially missing
data. Thus, we regard them as a suitable solution for detecting and appropriately representing anomalies and
their respective characteristics. In this position paper, we discuss the problem of automatic and interpretable
anomaly detection by means of GPs. That is, we elaborate on why GPs are well suited for anomaly detection
and what the current challenges are when applying these probabilistic models to large-scale production data.
1 INTRODUCTION
Anomaly detection is an important data mining pro-
cess to distinguish expected from unexpected data
patterns. It enables researchers as well as practitioners
to assess the condition and current state of a system
of interest (Chandola et al., 2009). Applications are
manifold, e.g. in medicine (Rajpurkar et al., 2017),
credit card fraud (Sorournejad et al., 2016), or net-
work intrusion detection (Ioannou et al., 2017). Es-
pecially with regards to industrial applications, moni-
toring sensor data from complex processes in order to
detect outliers or low-performing production behav-
ior caused by undesired patterns and trends, which
we summarize as anomalies, is a challenging task
(Beecks et al., 2019). Not only due to the massive
amount of sensor data but also due to different types
of anomalies, manual or automatic inspection systems
are frequently supported by anomaly detection algo-
rithms (Stojanovic et al., 2016). It is often stressed
that anomaly detection should be part of understand-
ing the production process as a whole (Niggemann
and Lohweg, 2015), i.e. that it is not sufficient to de-
tect anomalies but also to generate hypothesis as to
why they appear.
While the last years have witnessed the develop-
ment of different anomaly detection algorithms (cf.
the work of Renaudie et al. (2018) for a recent perfor-
mance evaluation in an industrial context) only less
effort has been spent on the investigation of the in-
herent structure of an anomaly. Utilizing techniques
of machine learning and Artificial Intelligence (AI)
for anomaly detection seems like a natural choice to
not only detect unexpected patterns, but to understand
them. Still, achieving the level of trustworthiness re-
quired for sensitive industrial applications is a non-
trivial task (cf. High-Level Expert Group on Artificial
Intelligence, 2019; Kwon et al., 2019).
With regards to the key requirements for trustwor-
thy AI defined by the High-Level Expert Group on
Artificial Intelligence (2019) of the European Com-
mission, we consider GPs (Rasmussen and Williams,
2006) as an appropriate machine learning model to
fulfill those demands. GPs deliver robust and reliable
statistical data models, which are able to cope with
unreliable, noisy, or partially missing data (Bo
ˇ
skoski
et al., 2012). Since these models are composed from
simple components defined by explainable hyperpa-
rameters, they are interpretable by nature and deliver
the capabilities to trace their predictions back to those
Berns, F., Lange-Hegermann, M. and Beecks, C.
Towards Gaussian Processes for Automatic and Interpretable Anomaly Detection in Industry 4.0.
DOI: 10.5220/0010130300870092
In Proceedings of the International Conference on Innovative Intelligent Industrial Production and Logistics (IN4PL 2020), pages 87-92
ISBN: 978-989-758-476-3
Copyright
c
2020 by SCITEPRESS Science and Technology Publications, Lda. All rights reserved
87
components and the according training data (Lloyd
et al., 2014; Duvenaud et al., 2013).
In this position paper, we discuss the problem
of automatic and interpretable anomaly detection by
means of GPs. That is, we elaborate on why GPs are
well suited for anomaly detection and what the cur-
rent challenges are when applying these probabilistic
models to large-scale production data. To this end,
our contributions are two-fold:
We introduce GPs as an interpretable model for
anomaly detection.
We outline challenges that have to be addressed in
order to scale GPs to large-scale production data
This position paper is structured as follows. We
outline related work in Section 2. We introduce GPs
in Section 3, while the concept and rationale for
anomaly detection by means of these statistical data
models and the remaining challenges in that field are
explained in detail in Section 4. We conclude our pa-
per with an outlook on future work in Section 5.
2 RELATED WORK
In the era of Industry 4.0, the field of anomaly de-
tection has become crucially important. As a re-
sult, there is a plethora of classical anomaly detec-
tion algorithms that have been proposed in recent
years such as Z-Score (Domingues et al., 2016), Ma-
halanobis Distance-Based, Empirical Covariance Es-
timation (Pedregosa et al., 2011; Chandola et al.,
2009), Robust Covariance Estimation (Rousseeuw,
1984; Chandola et al., 2009), Subspace-based PCA
Anomaly Detector (Chandola et al., 2009), One-Class
SVM (Sch
¨
olkopf et al., 2001; Pedregosa et al., 2011;
Chandola et al., 2009; Eskin et al., 2002), Isolation
Forest (I-Forest) (Liu et al., 2008; Pedregosa et al.,
2011), Gaussian Mixture Model (Pedregosa et al.,
2011; Chandola et al., 2009; Phua et al., 2010), Deep
Auto-Encoder (Candel et al., 2018; Gong et al., 2019),
Local Outlier Factor (Breunig et al., 2000; Pedregosa
et al., 2011; Chandola et al., 2009; Auslander et al.,
2011), Self-Organizing Maps (Von Birgelen et al.,
2018), Least Squares Anomaly Detector (Tavallaee
et al., 2010), GADPL (Graß et al., 2019), Automata
(Voden
ˇ
carevi
´
c et al., 2011), and k-Nearest Neighbor
(Goldstein and Uchida, 2016; Auslander et al., 2011;
Eskin et al., 2002).
Current approaches (Chalapathy and Chawla,
2019; Zenati et al., 2018; An and Cho, 2015; Zhang
and Chen, 2019; Sabokrou et al., 2018; Suh et al.,
2016; Berkhahn et al., 2019; Li et al., 2019; Kawachi
et al., 2018; Guo et al., 2018; Wang et al., 2019; Dias
et al., 2020) frequently make use of generative models
for anomaly detection, e.g. Variational Autoencoders
(Kingma and Welling, 2014), Generative Adversarial
Networks (Goodfellow et al., 2014), GP Latent Vari-
able Models (Damianou et al., 2016), or Normaliz-
ing Flows (Rezende and Mohamed, 2015), in partic-
ular for sequence data (Bowman et al., 2016). These
models can be trained automatically for the usage of
anomaly detection (M
¨
uller et al., 2020).
While these algorithms are all possible approaches
for anomaly detection, as shown in different surveys
(Goldstein and Uchida, 2016; Phua et al., 2010; Chan-
dola et al., 2009), they are not directly suited for de-
scribing the inherent structure of anomalies, which is
the major focus of this position paper. We choose
GPs (Rasmussen and Williams, 2006) for anomaly
description due to their capability to not only gather
statistical indicators, but deliver the very characteris-
tics of specific anomalous behavior from the data.
For automatically describing the underlying data
characteristics, Lloyd et al. (2014) have proposed the
Automatic Bayesian Covariance Discovery System
that adapts the Compositional Kernel Search Algo-
rithm (Duvenaud et al., 2013) by adding intuitive nat-
ural language descriptions of the function classes de-
scribed by their models. Hwang et al. (2016) fur-
ther expand on those concepts by expanding these
models to discover kernel structures which are able
to explain multiple time series at once. Recently,
the 3CS (Berns et al., 2020) and LARGe algorithms
(Berns and Beecks, 2020) have shown to outperform
the aforementioned approaches in terms of efficiency.
As these methods are all based on GPs, we give a
short introduction of GPs in the following section.
3 GAUSSIAN PROCESSES
A Gaussian Process (GP) (Rasmussen and Williams,
2006) is a stochastic process over random variables
{ f (x) | x X }, indexed by a set X , where every fi-
nite subset of random variables follows a multivariate
normal distribution. The distribution of a GP is the
joint distribution of all of these random variables and
it is thus a probability distribution over the space of
functions { f : X R}. A GP is formalized as
f (·) GP
m(·), k(·, ·)
, (1)
where the mean function m : X R and the covari-
ance function k : X ×X R are defined x, x
0
X via
m(x) = E [ f (x)] (2)
k(x, x
0
) = E
( f (x) m(x)) ·
f (x
0
) m(x
0
)

(3)
IN4PL 2020 - International Conference on Innovative Intelligent Industrial Production and Logistics
88
Given a finite dataset D = {X , Y } with input X =
{x
i
| x
i
X 1 i n} representing the underlying
index values, such as timestamps or locations, and tar-
get Y = {y
i
| y
i
= f (x
i
), x
i
X } representing the actual
data values, such as sensor values or other complex
measurements, a GP can be used to statistically repre-
sent the dataset D by optimizing the hyperparameters
θ of both mean and covariance function. This opti-
mization is frequently carried out by maximizing the
log marginal likelihood L (Rasmussen and Williams,
2006; Kim and Teh, 2018) of the GP:
L (m, k, θ | D) =
1
2
·
(y µ)
T
Σ
1
(y µ)+
log | Σ | + nlog(2π)] (4)
As can be seen in Equation 4, the marginalization
of a GP for a given dataset D of n records results in
a mean vector µ R
n
, and a covariance matrix Σ
R
n×n
which are defined as y[i] = f (x
i
), µ[i] = m(x
i
),
and Σ[i, j] = k(x
i
, x
j
) for 1 i, j n, respectively.
How GPs are utilized in the context of anomaly
detection and how the resulting challenges can be
solved are described in the following section.
4 GPs FOR ANOMALY
DETECTION
GP models are robust Bayesian models that intrinsi-
cally prevent overfitting. This sturdiness is a clear
advantage of GP models in dynamic and noisy pro-
duction environments with always changing circum-
stances like evolving processes or sensors modifica-
tions. GP models yield smooth models even when
only a few data points are given, which is particu-
lar important in production, where the batch size is
steadily decreasing.
In addition to recognizing an anomaly, it is impor-
tant to understand and interpret its underlying struc-
ture. Only this makes it possible to rectify the un-
derlying problem. Being interpretable is immanent in
GPs, due to their rigid mathematical structure. Their
covariance functions can induce combinations of var-
ious physically motivated behaviors like trends, pe-
riodicity, differential equations (Lange-Hegermann,
2018, 2020), and change points, and each of these
covariance functions comes with hyperparameters,
which are often interpretable physical constants like
periods. The model prediction can even be disas-
sembled into parts stemming from the individual in-
terpretable parts of the covariance function. All this
physically interpretable structure can be learned from
data, and additionally domain knowledge can be pre-
encoded into the models by experts.
What makes the application of GPs difficult to
large-scale, streaming production data is the super-
cubic computation time complexity of optimizing
GPs. Though efficient solutions exist (Snelson and
Ghahramani, 2007), they do not provide a fast solu-
tion for automatic and interpretable anomaly detec-
tion. For this reason, we identified the following ma-
jor challenges.
4.1 Efficient Algorithms for Large-scale
GP Models
The first challenge is concerned with the develop-
ment of efficient retrieval algorithms for GP models.
This includes the central question of how to efficiently
search and determine suitable covariance functions
for anomaly detection. Instead of applying a greedy
search through the space of possible covariance func-
tions, as done by state-of-the-art algorithms (Duve-
naud et al., 2013; Lloyd et al., 2014; Steinruecken
et al., 2019), one could learn the impact of individual
hyperparameters in order to specifically derive new
covariance functions. In addition to the development
of such intelligent covariance search heuristics, an-
other challenge is to process and infer GP models for
multivariate, event-based sensor data in (near) real
time. This demands for resource-efficient streaming
GP retrieval algorithms that scale to state of the art big
data processing frameworks such as Apache Hadoop,
Spark, and Storm.
4.2 Model Selection for GPs in Anomaly
Detection
Besides the development of efficient algorithms, the
second major challenge lies in the development of
suitable GP model selection approaches. While
prominent model quality estimators, such as the like-
lihood function, tend to prefer complex models con-
taining many hyperparameters, Laplace approxima-
tions enable to capture the model evidence of GP
models more properly. The particular challenge is
thus to couple Laplace approximations with unsuper-
vised GP Latent Variable Models (Damianou et al.,
2016) in order to enable conclusions on the underly-
ing physical processes and individual sensor dimen-
sions.
To sum up, the development of real-time GP
model retrieval and selection algorithms for multivari-
ate data streams that are open-source and are based
on open industry standards are necessary for detect-
ing, analyzing, and understanding anomalies in an
domain-agnostic, automatic, and interpretable man-
ner.
Towards Gaussian Processes for Automatic and Interpretable Anomaly Detection in Industry 4.0
89
5 CONCLUSION
In this paper, we have argued for the application of
GPs for automatic and interpretable anomaly detec-
tion. GPs are robust bayesian machine learning tools,
which enable inference for noisy as well as unreliable
data. GP models yield smooth models even when only
a few data points are given, which is particular useful
in industrial scenarios, where creating those records
is potentially expensive.
Along with the advantages, several challenges
have to be met. In particular, concepts and algorithms
are needed to facilitate efficient GP model retrieval
and selection on large-scale streaming data. We aim
to address these challenges in our future work and to
elaborate our developments in various application do-
mains.
REFERENCES
An, J. and Cho, S. (2015). Variational autoencoder based
anomaly detection using reconstruction probability.
Special Lecture on IE, 2(1):1–18.
Auslander, B., Gupta, K. M., and Aha, D. W. (2011). A
comparative evaluation of anomaly detection algo-
rithms for maritime video surveillance. In Carapezza,
E. M., editor, Proc. SPIE 8019, Sensors, and Com-
mand, Control, Communications, and Intelligence
(C3I) Technologies for Homeland Security and Home-
land Defense X, SPIE Proceedings, page 801907.
SPIE.
Beecks, C., Schmidt, K. W., Berns, F., and Graß, A. (2019).
Gaussian processes for anomaly description in pro-
duction environments. In EDBT/ICDT Workshops,
volume 2322 of CEUR Workshop Proceedings.
Berkhahn, F., Keys, R., Ouertani, W., Shetty, N., and
Geißler, D. (2019). Augmenting variational autoen-
coders with sparse labels: A unified framework for
unsupervised, semi-(un) supervised, and supervised
learning. arXiv preprint arXiv:1908.03015.
Berns, F. and Beecks, C. (2020). Automatic gaussian pro-
cess model retrieval for big data. In CIKM. ACM.
Berns, F., Schmidt, K., Bracht, I., and Beecks, C. (2020).
3CS Algorithm for Efficient Gaussian Process Model
Retrieval [accepted]. 25th International Conference
on Pattern Recognition (ICPR).
Bo
ˇ
skoski, P., Ga
ˇ
sperin, M., and Petelin, D. (2012). Bear-
ing fault prognostics based on signal complexity and
gaussian process models. In 2012 IEEE Conference
on Prognostics and Health Management, pages 1–8.
Bowman, S. R., Vilnis, L., Vinyals, O., Dai, A. M.,
J
´
ozefowicz, R., and Bengio, S. (2016). Generating
sentences from a continuous space. In CoNLL, pages
10–21. ACL.
Breunig, M., Kriegel, H.-P., Ng, R. T., and Sander, J.
(2000). Lof: Identifying density-based local outliers.
In Proceedings of the 2000 ACM SIGMOD Interna-
tional Conference on Management of Data, pages 93–
104. ACM.
Candel, A., LeDell, E., Parmar, V., and Arora,
A. (2018). Deep learning with h2o. https:
//www.h2o.ai/wp-content/themes/h2o2016/images/
resources/DeepLearningBooklet.pdf. (Accessed on
09/28/2020).
Chalapathy, R. and Chawla, S. (2019). Deep learning for
anomaly detection: A survey. CoRR, abs/1901.03407.
Chandola, V., Banerjee, A., and Kumar, V. (2009).
Anomaly detection: A survey. ACM computing sur-
veys (CSUR), 41(3):1–58.
Damianou, A. C., Titsias, M. K., and Lawrence, N. D.
(2016). Variational inference for latent variables and
uncertain inputs in gaussian processes. J. Mach.
Learn. Res., 17:42:1–42:62.
Dias, M. L. D., Mattos, C. L. C., da Silva, T. L. C.,
de Mac
ˆ
edo, J. A. F., and Silva, W. C. P. (2020).
Anomaly detection in trajectory data with normaliz-
ing flows. CoRR, abs/2004.05958.
Domingues, R., Buonora, F., Senesi, R., and Thonnard, O.
(2016). An application of unsupervised fraud detec-
tion to passenger name records. In 2016 46th Annual
IEEE/IFIP International Conference on Dependable
Systems and Networks Workshop (DSN-W), pages 54–
59.
Duvenaud, D., Lloyd, J. R., Grosse, R. B., Tenenbaum,
J. B., and Ghahramani, Z. (2013). Structure dis-
covery in nonparametric regression through composi-
tional kernel search. In ICML (3), volume 28 of JMLR
Workshop and Conference Proceedings, pages 1166–
1174. JMLR.org.
Eskin, E., Arnold, A., Prerau, M., Portnoy, L., and Stolfo,
S. (2002). A geometric framework for unsupervised
anomaly detection. In Barbar
´
a, D. and Jajodia, S., ed-
itors, Applications of Data Mining in Computer Secu-
rity, volume 6 of Advances in Information Security,
1568-2633, pages 77–101. Springer US and Imprint
and Springer, Boston, MA.
Goldstein, M. and Uchida, S. (2016). A comparative eval-
uation of unsupervised anomaly detection algorithms
for multivariate data. PloS one, 11(4):152–173.
Gong, D., Liu, L., Le, V., Saha, B., Mansour, M. R.,
Venkatesh, S., and Hengel, A. v. d. (2019). Memoriz-
ing normality to detect anomaly: Memory-augmented
deep autoencoder for unsupervised anomaly detec-
tion. In Proceedings of the IEEE International Con-
ference on Computer Vision, pages 1705–1714.
Goodfellow, I., Pouget-Abadie, J., Mirza, M., Xu, B.,
Warde-Farley, D., Ozair, S., Courville, A., and Ben-
gio, Y. (2014). Generative adversarial nets. In
NeurIPS.
Graß, A., Beecks, C., and Soto, J. A. C. (2019). Unsuper-
vised anomaly detection in production lines. In Bey-
erer, J., K
¨
uhnert, C., and Niggemann, O., editors, Ma-
chine Learning for Cyber Physical Systems, pages 18–
25, Berlin, Heidelberg. Springer Berlin Heidelberg.
Guo, Y., Liao, W., Wang, Q., Yu, L., Ji, T., and Li, P.
(2018). Multidimensional time series anomaly de-
IN4PL 2020 - International Conference on Innovative Intelligent Industrial Production and Logistics
90
tection: A gru-based gaussian mixture variational au-
toencoder approach. In Asian Conference on Machine
Learning, pages 97–112.
High-Level Expert Group on Artificial Intelligence
(2019). Ethics guidelines for trustworthy ai.
https://ec.europa.eu/digital-single-market/en/news/
ethics-guidelines-trustworthy-ai. (Accessed on
09/28/2020).
Hwang, Y., Tong, A., and Choi, J. (2016). Automatic con-
struction of nonparametric relational regression mod-
els for multiple time series. In Balcan, M. F. and Wein-
berger, K. Q., editors, ICML 2016: Proceedings of the
33rd International Conference on Machine Learning,
volume 48 of Proceedings of Machine Learning Re-
search, pages 3030–3039. PLMR.
Ioannou, C., Vassiliou, V., and Sergiou, C. (2017). An intru-
sion detection system for wireless sensor networks. In
2017 24th International Conference on Telecommuni-
cations (ICT), pages 1–5. IEEE.
Kawachi, Y., Koizumi, Y., and Harada, N. (2018). Com-
plementary set variational autoencoder for supervised
anomaly detection. In 2018 IEEE International Con-
ference on Acoustics, Speech and Signal Processing
(ICASSP), pages 2366–2370. IEEE.
Kim, H. and Teh, Y. W. (2018). Scaling up the Automatic
Statistician: Scalable structure discovery using Gaus-
sian processes. In Proceedings of the 21st Interna-
tional Conference on Artificial Intelligence and Statis-
tics, volume 84.
Kingma, D. P. and Welling, M. (2014). Auto-encoding vari-
ational bayes. In ICLR.
Kwon, D., Kim, H., Kim, J., Suh, S. C., Kim, I., and Kim,
K. J. (2019). A survey of deep learning-based network
anomaly detection. Cluster Computing, pages 1–13.
Lange-Hegermann, M. (2018). Algorithmic linearly con-
strained gaussian processes. In NeurIPS, pages 2141–
2152.
Lange-Hegermann, M. (2020). Linearly constrained gaus-
sian processes with boundary conditions. CoRR,
abs/2002.00818.
Li, D., Chen, D., Jin, B., Shi, L., Goh, J., and Ng, S.-K.
(2019). Mad-gan: Multivariate anomaly detection for
time series data with generative adversarial networks.
In International Conference on Artificial Neural Net-
works, pages 703–716. Springer.
Liu, F. T., Ting, K. M., and Zhou, Z.-H. (2008). Isolation
forest. In Giannotti, F., editor, Eighth IEEE Interna-
tional Conference on Data Mining, 2008, pages 413–
422, Piscataway, NJ. IEEE.
Lloyd, J. R., Duvenaud, D., Grosse, R. B., Tenenbaum,
J. B., and Ghahramani, Z. (2014). Automatic con-
struction and natural-language description of nonpara-
metric regression models. In AAAI, pages 1242–1250.
AAAI Press.
M
¨
uller, A., Lange-Hegermann, M., and von Birgelen,
A. (2020). Automatisches training eines varia-
tional autoencoder f
¨
ur anomalieerkennung in zeitrei-
hen. In VDI Kongress Automation 2020, volume VDI-
Berichte 2375, pages 687–698, Baden-Baden. VDI
Wissensforum GmbH, VDI Verlag GmbH.
Niggemann, O. and Lohweg, V. (2015). On the diagnosis of
cyber-physical production systems. In Twenty-Ninth
AAAI Conference on Artificial Intelligence.
Pedregosa, F., Varoquaux, G., Gramfort, A., Michel, V.,
Thirion, B., Grisel, O., Blondel, M., Prettenhofer, P.,
Weiss, R., Dubourg, V., Vanderplas, J., Passos, A.,
Cournapeau, D., Brucher, M., Perrot, M., and Duch-
esnay, E. (2011). Scikit-learn: Machine learning in
python. J. Mach. Learn. Res., 12:2825–2830.
Phua, C., Lee, V. C. S., Smith-Miles, K., and Gayler, R. W.
(2010). A comprehensive survey of data mining-based
fraud detection research. CoRR, abs/1009.6119.
Rajpurkar, P., Hannun, A. Y., Haghpanahi, M., Bourn, C.,
and Ng, A. Y. (2017). Cardiologist-level arrhythmia
detection with convolutional neural networks. CoRR,
abs/1707.01836.
Rasmussen, C. E. and Williams, C. K. I. (2006). Gaussian
Processes for Machine Learning (Adaptive Computa-
tion And Machine Learning). The MIT Press.
Renaudie, D., Zuluaga, M. A., and Acuna-Agost, R. (2018).
Benchmarking anomaly detection algorithms in an in-
dustrial context: Dealing with scarce labels and multi-
ple positive types. In IEEE International Conference
on Big Data, pages 1227–1236.
Rezende, D. J. and Mohamed, S. (2015). Variational infer-
ence with normalizing flows. In ICML, volume 37 of
JMLR Workshop and Conference Proceedings, pages
1530–1538. JMLR.org.
Rousseeuw, P. J. (1984). Least median of squares regres-
sion. Journal of the American statistical association,
79(388):871–880.
Sabokrou, M., Khalooei, M., Fathy, M., and Adeli, E.
(2018). Adversarially learned one-class classifier for
novelty detection. In Proceedings of the IEEE Con-
ference on Computer Vision and Pattern Recognition,
pages 3379–3388.
Sch
¨
olkopf, B., Platt, J. C., Shawe-Taylor, J. C., Smola, A. J.,
and Williamson, R. C. (2001). Estimating the support
of a high-dimensional distribution. Neural Comput.,
13(7):1443–1471.
Snelson, E. and Ghahramani, Z. (2007). Local and global
sparse gaussian process approximations. In AIS-
TATS, volume 2 of JMLR Proceedings, pages 524–
531. JMLR.org.
Sorournejad, S., Zojaji, Z., Atani, R. E., and Monadjemi,
A. H. (2016). A survey of credit card fraud detection
techniques: Data and technique oriented perspective.
CoRR, abs/1611.06439.
Steinruecken, C., Smith, E., Janz, D., Lloyd, J. R., and
Ghahramani, Z. (2019). The automatic statistician.
In Automated Machine Learning, The Springer Series
on Challenges in Machine Learning, pages 161–173.
Springer.
Stojanovic, L., Dinic, M., Stojanovic, N., and Stojadi-
novic, A. (2016). Big-data-driven anomaly detection
in industry (4.0): An approach and a case study. In
2016 IEEE International Conference on Big Data (Big
Data), pages 1647–1652.
Suh, S., Chae, D. H., Kang, H.-G., and Choi, S. (2016).
Echo-state conditional variational autoencoder for
Towards Gaussian Processes for Automatic and Interpretable Anomaly Detection in Industry 4.0
91
anomaly detection. In 2016 International Joint Con-
ference on Neural Networks (IJCNN), pages 1015–
1022. IEEE.
Tavallaee, M., Stakhanova, N., and Ghorbani, A. A.
(2010). Toward credible evaluation of anomaly-based
intrusion-detection methods. IEEE Transactions on
Systems, Man, and Cybernetics, Part C (Applications
and Reviews), 40(5):516–524.
Voden
ˇ
carevi
´
c, A., B
¨
uning, H. K., Niggemann, O., and
Maier, A. (2011). Using behavior models for anomaly
detection in hybrid systems. In 2011 XXIII Inter-
national Symposium on Information, Communication
and Automation Technologies, pages 1–8. IEEE.
Von Birgelen, A., Buratti, D., Mager, J., and Niggemann,
O. (2018). Self-organizing maps for anomaly local-
ization and predictive maintenance in cyber-physical
production systems. Procedia CIRP, 72:480–485.
Wang, X., Du, Y., Lin, S., Cui, P., and Yang, Y. (2019).
Self-adversarial variational autoencoder with gaus-
sian anomaly prior distribution for anomaly detection.
CoRR, abs/1903.00904.
Zenati, H., Romain, M., Foo, C.-S., Lecouat, B., and Chan-
drasekhar, V. (2018). Adversarially learned anomaly
detection. In 2018 IEEE International Conference on
Data Mining (ICDM), pages 727–736. IEEE.
Zhang, C. and Chen, Y. (2019). Time series anomaly
detection with variational autoencoders. CoRR,
abs/1907.01702.
IN4PL 2020 - International Conference on Innovative Intelligent Industrial Production and Logistics
92