Using wearable 360o camera, we collected a dataset of 360-degree egocentric videos of anomaly activities that visually impaired people commonly encounter in their daily lives.
We collected our dataset in various locations, including Automated Teller Machines (ATMs), Cafes, Bus Stations, and other public places. The dataset contains 276 normal and 299 abnormal videos.
Sample images of the three datasets, our VIEW360 dataset, UCF-Crime, and XD-Violence. This figure illustrates notable differences between our VIEW360 dataset and other widely used weakly-supervised anomaly detection datasets.
To overcome the limitations of existing weakly-supervised methods, which typically assign uniform anomaly scores to video snippets (a brief segment extracted from a video), leading to broad predictions, we developed the Frame and Direction Prediction Network (FDPN). FDPN enhances detection by analyzing anomalies at the frame level with a coarse-to-fine learning strategy, using snippet-level predictions for pseudo-supervision. This method is particularly effective for spotting sudden anomalies, like quick, unexpected events. Additionally, FDPN incorporates saliency-driven image masking to improve accuracy by focusing on visually prominent areas likely to contain anomalies. It also introduces direction classification for anomalies from an egocentric viewpoint, using a saliency heatmap to provide navigational assistance for visually impaired users.
The anomaly detection results of our FDPN on the VIEW360 dataset is illustrated in the graph, with ground truth abnormal frames are marked by blue box.
@InProceedings{Song_2025_WACV,
author = {Song, Inpyo and Lee, Sanghyeon and Joo, Minjun and Lee, Jangwon},
title = {Anomaly Detection for People with Visual Impairments using an Egocentric 360-Degree Camera},
booktitle = {Proceedings of the Winter Conference on Applications of Computer Vision (WACV)},
month = {February},
year = {2025},
pages = {2828-2837}
}