Track Any Object:
A Granular Video Anomaly Detection Pipeline

1Xiamen University 2The Chinese University of Hong Kong 3The University of Virginia
*Denotes Equal Contribution, ‡Denotes Corresponding Author
CVPR 2025

Abstract

Video anomaly detection (VAD) is crucial in scenarios such as surveillance and autonomous driving, where timely detection of unexpected activities is essential. Albeit existing methods have primarily focused on detecting anomalous objects in videos—either by identifying anomalous frames or objects—they often neglect finer-grained analysis, such as anomalous pixels, which limits their ability to capture a broader range of anomalies. To address this challenge, we propose an innovative VAD framework called Track Any Object (TAO), which introduces a Granular Video Anomaly Detection Framework that, for the first time, integrates the detection of multiple fine-grained anomalous objects into a unified framework. Unlike methods that assign anomaly scores to every pixel at each moment, our approach transforms the problem into pixel-level tracking of anomalous objects. By linking anomaly scores to subsequent tasks such as image segmentation and video tracking, our method eliminates the need for threshold selection and achieves more precise anomaly localization, even in longand challenging video sequences. Experiments on extensive datasets demonstrate that TAO achieves state-of-the-art performance, setting a new progress for VAD by providing a practical, granular, and holistic solution.

Existing VAD Models

Teaser Image

Video anomaly detection (VAD) models are predominantly framecentric or object-centric. Frame-centric methods detect anomalies in frames without localizing them, while object-centric methods identify anomalous objects but lack pixel-level accuracy. Pixel-centric models address these gaps by providing pixel-level localization, delivering fine-grained segmentation and precise delineation of anomalies, particularly for overlapping objects where traditional methods struggle.

Pipeline of TAO

Pipeline Image

(a) TAO first generate bounding boxes to identify objects in each frame. (b) Next, TAO score these boxes using object-centric video anomaly detection algorithms to extract potential anomalous boxes. To ensure robustness, TAO apply filtering to eliminate redundant boxes. (c) Finally, the filtered boxes and original frames are input into a prompt-based segmentation model to produce pixel-level anomaly segmentation masks.

Experimental Results

Result Image

Visual Results


Result Image

Result Image

Further Analysis


Result Image
Result Image

BibTeX