Show simple item record

dc.contributor.authorWu, Chongke
dc.contributor.authorShao, Sicong
dc.contributor.authorTunc, Cihan
dc.contributor.authorSatam, Pratik
dc.contributor.authorHariri, Salim
dc.date.accessioned2021-12-08T23:02:31Z
dc.date.available2021-12-08T23:02:31Z
dc.date.issued2021-11-23
dc.identifier.citationWu, C., Shao, S., Tunc, C., Satam, P., & Hariri, S. (2021). An explainable and efficient deep learning framework for video anomaly detection. Cluster Computing.en_US
dc.identifier.issn1386-7857
dc.identifier.doi10.1007/s10586-021-03439-5
dc.identifier.urihttp://hdl.handle.net/10150/662473
dc.description.abstractDeep learning-based video anomaly detection methods have drawn significant attention in the past few years due to their superior performance. However, almost all the leading methods for video anomaly detection rely on large-scale training datasets with long training times. As a result, many real-world video analysis tasks are still not applicable for fast deployment. On the other hand, the leading methods cannot provide interpretability due to the uninterpretable feature representations hiding the decision-making process when anomaly detection models are considered as a black box. However, the interpretability for anomaly detection is crucial since the corresponding response to the anomalies in the video is determined by their severity and nature. To tackle these problems, this paper proposes an efficient deep learning framework for video anomaly detection and provides explanations. The proposed framework uses pre-trained deep models to extract high-level concept and context features for training denoising autoencoder (DAE), requiring little training time (i.e., within 10 s on UCSD Pedestrian datasets) while achieving comparable detection performance to the leading methods. Furthermore, this framework presents the first video anomaly detection use of combing autoencoder and SHapley Additive exPlanations (SHAP) for model interpretability. The framework can explain each anomaly detection result in surveillance videos. In the experiments, we evaluate the proposed framework's effectiveness and efficiency while also explaining anomalies behind the autoencoder’s prediction. On the USCD Pedestrian datasets, the DAE achieved 85.9% AUC with a training time of 5 s on the USCD Ped1 and 92.4% AUC with a training time of 2.9 s on the UCSD Ped2.en_US
dc.description.sponsorshipAir Force Office of Scientific Researchen_US
dc.language.isoenen_US
dc.publisherSpringer Science and Business Media LLCen_US
dc.rights© The Author(s), under exclusive licence to Springer Science+Business Media, LLC, part of Springer Nature 2021.en_US
dc.rights.urihttp://rightsstatements.org/vocab/InC/1.0/en_US
dc.subjectAbnormal event detectionen_US
dc.subjectAnomaly video analysisen_US
dc.subjectContext miningen_US
dc.subjectDeep featuresen_US
dc.subjectInterpretabilityen_US
dc.subjectSecurityen_US
dc.subjectVideo surveillanceen_US
dc.titleAn explainable and efficient deep learning framework for video anomaly detectionen_US
dc.typeArticleen_US
dc.identifier.eissn1573-7543
dc.contributor.departmentNSF Center for Cloud and Autonomic Computing, The University of Arizonaen_US
dc.identifier.journalCluster Computingen_US
dc.description.note12 month embargo; published: 23 November 2021en_US
dc.description.collectioninformationThis item from the UA Faculty Publications collection is made available by the University of Arizona with support from the University of Arizona Libraries. If you have questions, please contact us at repository@u.library.arizona.edu.en_US
dc.eprint.versionFinal accepted manuscripten_US
dc.identifier.pii3439
dc.source.journaltitleCluster Computing


Files in this item

Thumbnail
Name:
Explain_VAD_v53_cam.pdf
Size:
1.293Mb
Format:
PDF
Description:
Final Accepted Manuscript

This item appears in the following Collection(s)

Show simple item record