Diffusion model
updated
Your Student is Better Than Expected: Adaptive Teacher-Student
Collaboration for Text-Conditional Diffusion Models
Paper
• 2312.10835
• Published
• 7
LIME: Localized Image Editing via Attention Regularization in Diffusion
Models
Paper
• 2312.09256
• Published
• 10
PromptBench: A Unified Library for Evaluation of Large Language Models
Paper
• 2312.07910
• Published
• 16
Prompt Expansion for Adaptive Text-to-Image Generation
Paper
• 2312.16720
• Published
• 6
A Recipe for Scaling up Text-to-Video Generation with Text-free Videos
Paper
• 2312.15770
• Published
• 15
DreamTuner: Single Image is Enough for Subject-Driven Generation
Paper
• 2312.13691
• Published
• 27
AnyV2V: A Plug-and-Play Framework For Any Video-to-Video Editing Tasks
Paper
• 2403.14468
• Published
• 27
HyperLLaVA: Dynamic Visual and Language Expert Tuning for Multimodal
Large Language Models
Paper
• 2403.13447
• Published
• 19
AnimateDiff-Lightning: Cross-Model Diffusion Distillation
Paper
• 2403.12706
• Published
• 18
FouriScale: A Frequency Perspective on Training-Free High-Resolution
Image Synthesis
Paper
• 2403.12963
• Published
• 8
LightIt: Illumination Modeling and Control for Diffusion Models
Paper
• 2403.10615
• Published
• 18
DepthFM: Fast Monocular Depth Estimation with Flow Matching
Paper
• 2403.13788
• Published
• 18
Video Editing via Factorized Diffusion Distillation
Paper
• 2403.09334
• Published
• 22
ELLA: Equip Diffusion Models with LLM for Enhanced Semantic Alignment
Paper
• 2403.05135
• Published
• 45
AtomoVideo: High Fidelity Image-to-Video Generation
Paper
• 2403.01800
• Published
• 23
VideoElevator: Elevating Video Generation Quality with Versatile
Text-to-Image Diffusion Models
Paper
• 2403.05438
• Published
• 20
DiffuseKronA: A Parameter Efficient Fine-tuning Method for Personalized
Diffusion Model
Paper
• 2402.17412
• Published
• 23
Playground v2.5: Three Insights towards Enhancing Aesthetic Quality in
Text-to-Image Generation
Paper
• 2402.17245
• Published
• 11
RealCompo: Dynamic Equilibrium between Realism and Compositionality
Improves Text-to-Image Diffusion Models
Paper
• 2402.12908
• Published
• 10
Seeing and Hearing: Open-domain Visual-Audio Generation with Diffusion
Latent Aligners
Paper
• 2402.17723
• Published
• 16
Getting it Right: Improving Spatial Consistency in Text-to-Image Models
Paper
• 2404.01197
• Published
• 31
StoryDiffusion: Consistent Self-Attention for Long-Range Image and Video
Generation
Paper
• 2405.01434
• Published
• 56
Paint by Inpaint: Learning to Add Image Objects by Removing Them First
Paper
• 2404.18212
• Published
• 30
Customizing Text-to-Image Models with a Single Image Pair
Paper
• 2405.01536
• Published
• 22