RefEdit: A Benchmark and Method for Improving Instruction-based Image Editing Model on Referring Expressions Paper • 2506.03448 • Published 3 days ago • 4
RelationAdapter: Learning and Transferring Visual Relation with Diffusion Transformers Paper • 2506.02528 • Published 4 days ago • 15
DCM: Dual-Expert Consistency Model for Efficient and High-Quality Video Generation Paper • 2506.03123 • Published 3 days ago • 14
FlowMo: Variance-Based Flow Guidance for Coherent Motion in Video Generation Paper • 2506.01144 • Published 5 days ago • 14
Sparse-vDiT: Unleashing the Power of Sparse Attention to Accelerate Video Diffusion Transformers Paper • 2506.03065 • Published 3 days ago • 27
RenderFormer: Transformer-based Neural Rendering of Triangle Meshes with Global Illumination Paper • 2505.21925 • Published 10 days ago • 34
Chain-of-Zoom: Extreme Super-Resolution via Scale Autoregression and Preference Alignment Paper • 2505.18600 • Published 13 days ago • 45
Temporal In-Context Fine-Tuning for Versatile Control of Video Diffusion Models Paper • 2506.00996 • Published 5 days ago • 34
Cora: Correspondence-aware image editing using few step diffusion Paper • 2505.23907 • Published 8 days ago • 11
MakeAnything: Harnessing Diffusion Transformers for Multi-Domain Procedural Sequence Generation Paper • 2502.01572 • Published Feb 3 • 21
UniGeo: Taming Video Diffusion for Unified Consistent Geometry Estimation Paper • 2505.24521 • Published 7 days ago • 15
Alchemist: Turning Public Text-to-Image Data into Generative Gold Paper • 2505.19297 • Published 12 days ago • 73
Memory-Efficient Visual Autoregressive Modeling with Scale-Aware KV Cache Compression Paper • 2505.19602 • Published 11 days ago • 13
Direct3D-S2: Gigascale 3D Generation Made Easy with Spatial Sparse Attention Paper • 2505.17412 • Published 15 days ago • 18
Model Already Knows the Best Noise: Bayesian Active Noise Selection via Attention in Video Diffusion Model Paper • 2505.17561 • Published 14 days ago • 30