metadata
license: apache-2.0
Visual Anomaly Detection under Complex View-Illumination Interplay: A Large-Scale Benchmark
π Hugging Face Dataset
π Paper β’ π Homepage
by Yunkang Cao*, Yuqi Cheng*, Xiaohao Xu, Yiheng Zhang, Yihan Sun, Yuxiang Tan, Yuxin Zhang, Weiming Shen,
π Updates
We're committed to open science! Here's our progress:
- 2025/05/19: π Paper released on ArXiv.
- 2025/05/16: π Dataset homepage launched.
- 2025/05/24: π§ͺ Code release for benchmark evaluation! code
π Introduction
Visual Anomaly Detection (VAD) systems often fail in the real world due to sensitivity to viewpoint-illumination interplayβcomplex interactions that distort defect visibility. Existing benchmarks overlook this challenge.
Introducing M2AD (Multi-View Multi-Illumination Anomaly Detection), a large-scale benchmark designed to rigorously test VAD robustness under these conditions:
- 119,880 high-resolution images across 10 categories, 999 specimens, 12 views, and 10 illuminations (120 configurations).
- Two evaluation protocols:
- π M2AD-Synergy: Tests multi-configuration information fusion.
- π§ͺ M2AD-Invariant: Measures single-image robustness to view-illumination variations.
- Key finding: SOTA VAD methods struggle significantly on M2AD, highlighting the critical need for robust solutions.