-
Frame Guidance: Training-Free Guidance for Frame-Level Control in Video Diffusion Models
Paper • 2506.07177 • Published • 22 -
StyleMaster: Stylize Your Video with Artistic Generation and Translation
Paper • 2412.07744 • Published • 20 -
ColoristaNet for Photorealistic Video Style Transfer
Paper • 2212.09247 • Published -
Style-A-Video: Agile Diffusion for Arbitrary Text-based Video Style Transfer
Paper • 2305.05464 • Published
Collections
Discover the best community collections!
Collections including paper arxiv:2403.14468
-
Divide-or-Conquer? Which Part Should You Distill Your LLM?
Paper • 2402.15000 • Published • 25 -
Adding NVMe SSDs to Enable and Accelerate 100B Model Fine-tuning on a Single GPU
Paper • 2403.06504 • Published • 55 -
MoAI: Mixture of All Intelligence for Large Language and Vision Models
Paper • 2403.07508 • Published • 77 -
AnyV2V: A Plug-and-Play Framework For Any Video-to-Video Editing Tasks
Paper • 2403.14468 • Published • 27
-
Your Student is Better Than Expected: Adaptive Teacher-Student Collaboration for Text-Conditional Diffusion Models
Paper • 2312.10835 • Published • 7 -
LIME: Localized Image Editing via Attention Regularization in Diffusion Models
Paper • 2312.09256 • Published • 12 -
PromptBench: A Unified Library for Evaluation of Large Language Models
Paper • 2312.07910 • Published • 19 -
Prompt Expansion for Adaptive Text-to-Image Generation
Paper • 2312.16720 • Published • 6
-
MagicVideo-V2: Multi-Stage High-Aesthetic Video Generation
Paper • 2401.04468 • Published • 50 -
Anything in Any Scene: Photorealistic Video Object Insertion
Paper • 2401.17509 • Published • 17 -
Memory Consolidation Enables Long-Context Video Understanding
Paper • 2402.05861 • Published • 10 -
Magic-Me: Identity-Specific Video Customized Diffusion
Paper • 2402.09368 • Published • 31
-
Can Large Language Models Understand Context?
Paper • 2402.00858 • Published • 24 -
OLMo: Accelerating the Science of Language Models
Paper • 2402.00838 • Published • 84 -
Self-Rewarding Language Models
Paper • 2401.10020 • Published • 152 -
SemScore: Automated Evaluation of Instruction-Tuned LLMs based on Semantic Textual Similarity
Paper • 2401.17072 • Published • 24
-
High-Quality Image Restoration Following Human Instructions
Paper • 2401.16468 • Published • 14 -
Object-Driven One-Shot Fine-tuning of Text-to-Image Diffusion with Prototypical Embedding
Paper • 2401.15708 • Published • 12 -
Taiyi-Diffusion-XL: Advancing Bilingual Text-to-Image Generation with Large Vision-Language Model Support
Paper • 2401.14688 • Published • 13 -
TIP-Editor: An Accurate 3D Editor Following Both Text-Prompts And Image-Prompts
Paper • 2401.14828 • Published • 9
-
One-for-All: Generalized LoRA for Parameter-Efficient Fine-tuning
Paper • 2306.07967 • Published • 24 -
Rerender A Video: Zero-Shot Text-Guided Video-to-Video Translation
Paper • 2306.07954 • Published • 111 -
TryOnDiffusion: A Tale of Two UNets
Paper • 2306.08276 • Published • 74 -
Seeing the World through Your Eyes
Paper • 2306.09348 • Published • 33
-
Frame Guidance: Training-Free Guidance for Frame-Level Control in Video Diffusion Models
Paper • 2506.07177 • Published • 22 -
StyleMaster: Stylize Your Video with Artistic Generation and Translation
Paper • 2412.07744 • Published • 20 -
ColoristaNet for Photorealistic Video Style Transfer
Paper • 2212.09247 • Published -
Style-A-Video: Agile Diffusion for Arbitrary Text-based Video Style Transfer
Paper • 2305.05464 • Published
-
Divide-or-Conquer? Which Part Should You Distill Your LLM?
Paper • 2402.15000 • Published • 25 -
Adding NVMe SSDs to Enable and Accelerate 100B Model Fine-tuning on a Single GPU
Paper • 2403.06504 • Published • 55 -
MoAI: Mixture of All Intelligence for Large Language and Vision Models
Paper • 2403.07508 • Published • 77 -
AnyV2V: A Plug-and-Play Framework For Any Video-to-Video Editing Tasks
Paper • 2403.14468 • Published • 27
-
Can Large Language Models Understand Context?
Paper • 2402.00858 • Published • 24 -
OLMo: Accelerating the Science of Language Models
Paper • 2402.00838 • Published • 84 -
Self-Rewarding Language Models
Paper • 2401.10020 • Published • 152 -
SemScore: Automated Evaluation of Instruction-Tuned LLMs based on Semantic Textual Similarity
Paper • 2401.17072 • Published • 24
-
Your Student is Better Than Expected: Adaptive Teacher-Student Collaboration for Text-Conditional Diffusion Models
Paper • 2312.10835 • Published • 7 -
LIME: Localized Image Editing via Attention Regularization in Diffusion Models
Paper • 2312.09256 • Published • 12 -
PromptBench: A Unified Library for Evaluation of Large Language Models
Paper • 2312.07910 • Published • 19 -
Prompt Expansion for Adaptive Text-to-Image Generation
Paper • 2312.16720 • Published • 6
-
High-Quality Image Restoration Following Human Instructions
Paper • 2401.16468 • Published • 14 -
Object-Driven One-Shot Fine-tuning of Text-to-Image Diffusion with Prototypical Embedding
Paper • 2401.15708 • Published • 12 -
Taiyi-Diffusion-XL: Advancing Bilingual Text-to-Image Generation with Large Vision-Language Model Support
Paper • 2401.14688 • Published • 13 -
TIP-Editor: An Accurate 3D Editor Following Both Text-Prompts And Image-Prompts
Paper • 2401.14828 • Published • 9
-
MagicVideo-V2: Multi-Stage High-Aesthetic Video Generation
Paper • 2401.04468 • Published • 50 -
Anything in Any Scene: Photorealistic Video Object Insertion
Paper • 2401.17509 • Published • 17 -
Memory Consolidation Enables Long-Context Video Understanding
Paper • 2402.05861 • Published • 10 -
Magic-Me: Identity-Specific Video Customized Diffusion
Paper • 2402.09368 • Published • 31
-
One-for-All: Generalized LoRA for Parameter-Efficient Fine-tuning
Paper • 2306.07967 • Published • 24 -
Rerender A Video: Zero-Shot Text-Guided Video-to-Video Translation
Paper • 2306.07954 • Published • 111 -
TryOnDiffusion: A Tale of Two UNets
Paper • 2306.08276 • Published • 74 -
Seeing the World through Your Eyes
Paper • 2306.09348 • Published • 33