Anomaly Detection Using Spatio-Temporal Context Learned by Video Clip Sorting

published in IEICE Transactions on Information and Systems

2022 Volume E105.D Issue 5 Pages 1094-1102

Wen SHAO, Rei KAWAKAMI, Takeshi NAEMURA

Previous studies on anomaly detection in videos have trained detectors in which reconstruction and prediction tasks are performed on normal data so that frames on which their task performance is low will be detected as anomalies during testing. This paper proposes a new approach that involves sorting video clips, by using a generative network structure. Our approach learns spatial contexts from appearances and temporal contexts from the order relationship of the frames. Experiments were conducted on four datasets, and we categorized the anomalous sequences by appearance and motion. Evaluations were conducted not only on each total dataset but also on each of the categories. Our method improved detection performance on both anomalies with different appearance and different motion from normality. Moreover, combining our approach with a prediction method produced improvements in precision at a high recall.


Paper
Anomalous sequences category annotation