Unbiased multiple instance learning for weakly supervised video anomaly detection
Weakly Supervised Video Anomaly Detection (WSVAD) is challenging because the binary anomaly label is only given on the video level, but the output requires snippetlevel predictions. So, Multiple Instance Learning (MIL) is prevailing in WSVAD. However, MIL is notoriously known to suffer from man...
Saved in:
Main Authors: | , , , , , |
---|---|
Other Authors: | |
Format: | Conference or Workshop Item |
Language: | English |
Published: |
2023
|
Subjects: | |
Online Access: | https://hdl.handle.net/10356/169509 https://cvpr2023.thecvf.com/Conferences/2023 |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Institution: | Nanyang Technological University |
Language: | English |
Summary: | Weakly Supervised Video Anomaly Detection (WSVAD)
is challenging because the binary anomaly label is only
given on the video level, but the output requires snippetlevel
predictions. So, Multiple Instance Learning (MIL) is
prevailing in WSVAD. However, MIL is notoriously known
to suffer from many false alarms because the snippet-level
detector is easily biased towards the abnormal snippets with
simple context, confused by the normality with the same
bias, and missing the anomaly with a different pattern. To
this end, we propose a new MIL framework: Unbiased MIL
(UMIL), to learn unbiased anomaly features that improve
WSVAD. At each MIL training iteration, we use the current
detector to divide the samples into two groups with different
context biases: the most confident abnormal/normal
snippets and the rest ambiguous ones. Then, by seeking
the invariant features across the two sample groups, we
can remove the variant context biases. Extensive experiments
on benchmarks UCF-Crime and TAD demonstrate
the effectiveness of our UMIL. Our code is provided at
https://github.com/ktr-hubrt/UMIL. |
---|