Abstract
Autoencoders have become increasingly popular in anomaly detection tasks over the years. Nevertheless, it remains a challenge to train autoencoders for anomaly detection tasks properly. A key contributing factor to this problem in many applications is the absence of a clean dataset from which the normal case can be learned. Instead, autoencoders must be trained based on a contaminated dataset containing an unknown amount of anomalies that potentially harm the training process. In this paper, we address this problem by studying the impact of the loss function on the robustness of an autoencoder. It is common practice to train an autoencoder by minimizing a loss function (e.g. squared error loss) under the assumption that all features are equally important to be reconstructed well. We relax this assumption and introduce a new loss function that adapts its robustness to anomalies based on the characteristics of data and on a per feature basis. Experimental results show that an autoencoder can be trained by this loss function robustly even when the training process is subject to many anomalies.
Original language | English |
---|---|
Title of host publication | Proceedings of the 22nd Engineering Applications of Neural Networks Conference |
Subtitle of host publication | EANN 2021 |
Editors | Lazaros Iliadis, John MacIntyre, Chrisina Jayne, Elias Pimenidis |
Publisher | Springer, Cham |
Pages | 42-53 |
Number of pages | 11 |
ISBN (Electronic) | 978-3-030-80568-5 |
ISBN (Print) | 978-3-030-80567-8 |
DOIs | |
Publication status | Published - 1 Jul 2021 |
Event | 22nd Engineering Applications of Neural Networks Conference - remote LIVE conference Duration: 25 Jun 2021 → 27 Jun 2021 http://www.eann2021.eu/ |
Publication series
Series | Proceedings of the International Neural Networks Society |
---|---|
Volume | 3 |
Conference
Conference | 22nd Engineering Applications of Neural Networks Conference |
---|---|
Abbreviated title | EANN 2021 |
Period | 25/06/21 → 27/06/21 |
Internet address |