AD Research Hub
DashboardPeoplePapersWorkshopsDatasetsResearch Map
Login
⌘K
AD Research Hub — Anomaly Detection in Computer Vision
← Back to Papers
Just Dance with pi! A Poly-modal Inductor for Weakly-supervised Video Anomaly Detection

Authors

Snehashis Majhi, Giacomo D'Amicantonio, Antitza Dantcheva, Quan Kong, Lorenzo Garattoni, Gianpiero Francesca, Egor Bondarev, Francois Bremond

CVPR-2025direct anomaly

Score

24

Tags

anomaly detectionvideo anomaly

Links

Paper PagearXiv AbstractarXiv PDF

Cite

Related Papers

Track Any Anomalous Object:A Granular Video Anomaly Detection Pipeline

Yuzhi Huang, Chenxin Li, Haitao Zhang +9

CVPR-2025direct anomaly34
anomaly detectionanomalousvideo anomaly
PDFarXiv

Anomize: Better Open Vocabulary Video Anomaly Detection

Fei Li, Wenxuan Liu, Jingjing Chen +4

CVPR-2025direct anomaly24
anomaly detectionvideo anomaly
PDFarXiv

Noise-Resistant Video Anomaly Detection via RGB Error-Guided Multiscale Predictive Coding and Dynamic Memory

Han Hu, Wenli Du, Peng Liao +2

CVPR-2025direct anomaly24
anomaly detectionvideo anomaly
PDF

VERA: Explainable Video Anomaly Detection via Verbalized Learning of Vision-Language Models

Muchao Ye, Weiyang Liu, Pan He

CVPR-2025direct anomaly24
anomaly detectionvideo anomaly
PDFarXiv

Weakly Supervised Video Anomaly Detection with Anomaly-Connected Components and Intention Reasoning

Yu Wang, Shengjie Zhao

Weakly supervised video anomaly detection (WS-VAD) involves identifying the temporal intervals that contain anomalous events in untrimmed videos, where only video-level annotations are provided as supervisory signals. However, a key limitation persists in WS-VAD, as dense frame-level annotations are absent, which often leaves existing methods struggling to learn anomaly semantics effectively. To address this issue, we propose a novel framework named LAS-VAD, short for Learning Anomaly Semantics for WS-VAD, which integrates anomaly-connected component mechanism and intention awareness mechanism

CVPR-2026direct anomaly40
anomaly detectionanomalousabnormal behaviorvideo anomaly+2
PDFarXiv

No Need For Real Anomaly: MLLM Empowered Zero-Shot Video Anomaly Detection

Zunkai Dai, Ke Li, Jiajia Liu +2

The collection and detection of video anomaly data has long been a challenging problem due to its rare occurrence and spatio-temporal scarcity. Existing video anomaly detection (VAD) methods under perform in open-world scenarios. Key contributing factors include limited dataset diversity, and inadequate understanding of context-dependent anomalous semantics. To address these issues, i) we propose LAVIDA, an end-to-end zero-shot video anomaly detection framework. ii) LAVIDA employs an Anomaly Exposure Sampler that transforms segmented objects into pseudo-anomalies to enhance model adaptability

CVPR-2026direct anomaly30
anomaly detectionanomalousvideo anomalyarxiv+1
ViTZero-shot
PDFarXiv

Sign in to access this content

Sign in