TY - JOUR
T1 - Multivariate anomaly detection based on prediction intervals constructed using deep learning
AU - Mathonsi, Thabang
AU - Zyl, Terence L.van
N1 - Publisher Copyright:
© The Author(s), under exclusive licence to Springer-Verlag London Ltd., part of Springer Nature 2021.
PY - 2025/1
Y1 - 2025/1
N2 - It has been shown that deep learning models can under certain circumstances outperform traditional statistical methods at forecasting. Furthermore, various techniques have been developed for quantifying the forecast uncertainty (prediction intervals). In this paper, we utilize prediction intervals constructed with the aid of artificial neural networks to detect anomalies in the multivariate setting. Challenges with existing deep learning-based anomaly detection approaches include (i) large sets of parameters that may be computationally intensive to tune, (ii) returning too many false positives rendering the techniques impractical for use, and (iii) requiring labeled datasets for training which are often not prevalent in real life. Our approach overcomes these challenges. We benchmark our approach against the oft-preferred well-established statistical models. We focus on three deep learning architectures, namely cascaded neural networks, reservoir computing, and long short-term memory recurrent neural networks. Our finding is deep learning outperforms (or at the very least is competitive to) the latter.
AB - It has been shown that deep learning models can under certain circumstances outperform traditional statistical methods at forecasting. Furthermore, various techniques have been developed for quantifying the forecast uncertainty (prediction intervals). In this paper, we utilize prediction intervals constructed with the aid of artificial neural networks to detect anomalies in the multivariate setting. Challenges with existing deep learning-based anomaly detection approaches include (i) large sets of parameters that may be computationally intensive to tune, (ii) returning too many false positives rendering the techniques impractical for use, and (iii) requiring labeled datasets for training which are often not prevalent in real life. Our approach overcomes these challenges. We benchmark our approach against the oft-preferred well-established statistical models. We focus on three deep learning architectures, namely cascaded neural networks, reservoir computing, and long short-term memory recurrent neural networks. Our finding is deep learning outperforms (or at the very least is competitive to) the latter.
KW - Anomaly detection
KW - Deep learning
KW - Multivariate time series forecasting
KW - Prediction intervals
UR - http://www.scopus.com/inward/record.url?scp=85122239004&partnerID=8YFLogxK
U2 - 10.1007/s00521-021-06697-x
DO - 10.1007/s00521-021-06697-x
M3 - Article
AN - SCOPUS:85122239004
SN - 0941-0643
VL - 37
SP - 707
EP - 721
JO - Neural Computing and Applications
JF - Neural Computing and Applications
IS - 2
ER -