Improving Zero-Shot Object-Level Change Detection by Incorporating Visual Correspondence Paper • 2501.05555 • Published Jan 9 • 1
TAB: Transformer Attention Bottlenecks enable User Intervention and Debugging in Vision-Language Models Paper • 2412.18675 • Published Dec 24, 2024 • 1
B-score: Detecting biases in large language models using response history Paper • 2505.18545 • Published May 24 • 31
VideoGameQA-Bench: Evaluating Vision-Language Models for Video Game Quality Assurance Paper • 2505.15952 • Published May 21 • 20
Understanding Generative AI Capabilities in Everyday Image Editing Tasks Paper • 2505.16181 • Published May 22 • 24
HoT: Highlighted Chain of Thought for Referencing Supporting Facts from Inputs Paper • 2503.02003 • Published Mar 3 • 48
ZeroBench: An Impossible Visual Benchmark for Contemporary Large Multimodal Models Paper • 2502.09696 • Published Feb 13 • 44
VideoGameBunny: Towards vision assistants for video games Paper • 2407.15295 • Published Jul 21, 2024 • 22
Zoom is what you need: An empirical study of the power of zoom and spatial biases in image classification Paper • 2304.05538 • Published Apr 11, 2023 • 2