Inference-Time Scaling for Diffusion Models beyond Scaling Denoising Steps

Type: Preprint

Publication Date: 2025-01-16

Citations: 0

DOI: https://doi.org/10.48550/arxiv.2501.09732

Abstract

Generative models have made significant impacts across various domains, largely due to their ability to scale during training by increasing data, computational resources, and model size, a phenomenon characterized by the scaling laws. Recent research has begun to explore inference-time scaling behavior in Large Language Models (LLMs), revealing how performance can further improve with additional computation during inference. Unlike LLMs, diffusion models inherently possess the flexibility to adjust inference-time computation via the number of denoising steps, although the performance gains typically flatten after a few dozen. In this work, we explore the inference-time scaling behavior of diffusion models beyond increasing denoising steps and investigate how the generation performance can further improve with increased computation. Specifically, we consider a search problem aimed at identifying better noises for the diffusion sampling process. We structure the design space along two axes: the verifiers used to provide feedback, and the algorithms used to find better noise candidates. Through extensive experiments on class-conditioned and text-conditioned image generation benchmarks, our findings reveal that increasing inference-time compute leads to substantial improvements in the quality of samples generated by diffusion models, and with the complicated nature of images, combinations of the components in the framework can be specifically chosen to conform with different application scenario.

Locations

  • arXiv (Cornell University) - View - PDF

Similar Works

Action Title Year Authors
+ PDF Chat Fast Solvers for Discrete Diffusion Models: Theory and Applications of High-Order Algorithms 2025 Yinuo Ren
Haoxuan Chen
Yuchen Zhu
Wei-Feng Guo
Yongxin Chen
Grant M. Rotskoff
Molei Tao
Lexing Ying
+ PDF Chat EM Distillation for One-step Diffusion Models 2024 Sirui Xie
Zhisheng Xiao
Diederik P. Kingma
Tingbo Hou
Ying Wu
Kevin Murphy
Tim Salimans
Ben Poole
Ruiqi Gao
+ PDF Chat Bigger is not Always Better: Scaling Properties of Latent Diffusion Models 2024 Kangfu Mei
Zhengzhong Tu
Mauricio Delbracio
Hossein Talebi
Vishal M. Patel
Peyman Milanfar
+ PDF Chat LD-Pruner: Efficient Pruning of Latent Diffusion Models using Task-Agnostic Insights 2024 Thibault Castells
Hyoung‐Kyu Song
Bo-Kyeong Kim
Shinkook Choi
+ Fast Sampling via De-randomization for Discrete Diffusion Models 2023 Zixiang Chen
Huizhuo Yuan
Yongqian Li
Yiwen Kou
Junkai Zhang
Quanquan Gu
+ PDF Chat Denoising Task Difficulty-based Curriculum for Training Diffusion Models 2024 Jin Young Kim
Hyojun Go
Soonwoo Kwon
Hyun‐Gyoon Kim
+ Diffusion Models: A Comprehensive Survey of Methods and Applications 2022 L. Yang
Zhilong Zhang
Shenda Hong
+ PDF Chat Scaling up Masked Diffusion Models on Text 2024 Shen Nie
Fengqi Zhu
Chao Du
Tianyu Pang
Qian Liu
Guihua Zeng
Min Lin
Chongxuan Li
+ PDF Chat Plug-and-Play Diffusion Distillation 2024 Yi-Ting Hsiao
Siavash Khodadadeh
Kevin Duarte
Wei-An Lin
Hui Qu
Mingi Kwon
Ratheesh Kalarot
+ PDF Chat PQD: Post-training Quantization for Efficient Diffusion Models 2024 Jiaojiao Ye
Zhen Wang
Jiang Li
+ Your Student is Better Than Expected: Adaptive Teacher-Student Collaboration for Text-Conditional Diffusion Models 2023 Nikita Starodubcev
Artem Fedorov
Artem Babenko
Dmitry Baranchuk
+ BOOT: Data-free Distillation of Denoising Diffusion Models with Bootstrapping 2023 Jiatao Gu
Shuangfei Zhai
Yizhe Zhang
Lingjie Liu
Josh Susskind
+ Diffusion Models: A Comprehensive Survey of Methods and Applications 2023 L. Yang
Zhilong Zhang
Yang Song
Shenda Hong
Runsheng Xu
Yue Zhao
Wentao Zhang
Bin Cui
Ming–Hsuan Yang
+ Faster Diffusion: Rethinking the Role of UNet Encoder in Diffusion Models 2023 Senmao Li
Taihang Hu
Fahad Shahbaz Khan
Linxuan Li
Shiqi Yang
Yaxing Wang
Ming‐Ming Cheng
Jian Yang
+ PDF Chat Text Diffusion with Reinforced Conditioning 2024 Yuxuan Liu
Tianchi Yang
Shaohan Huang
Zihan Zhang
Haizhen Huang
Furu Wei
Weiwei Deng
Feng Sun
Qi Zhang
+ PDF Chat Text Diffusion with Reinforced Conditioning 2024 Yuxuan Liu
Tianchi Yang
Shaohan Huang
Zihan Zhang
Haizhen Huang
Furu Wei
Weiwei Deng
Feng Sun
Qi Zhang
+ Self-conditioned Embedding Diffusion for Text Generation 2022 Robin Strudel
Corentin Tallec
Florent AltchĂŠ
Yilun Du
Yaroslav Ganin
Arthur Mensch
Will Grathwohl
Nikolay Savinov
Sander Dieleman
Laurent Sifre
+ PDF Chat FlashEval: Towards Fast and Accurate Evaluation of Text-to-image Diffusion Generative Models 2024 Lin Zhao
Tianchen Zhao
Zinan Lin
Xuefei Ning
Guohao Dai
Huazhong Yang
Yu Wang
+ Understanding Diffusion Models: A Unified Perspective 2022 Calvin Luo
+ PDF Chat $\textit{Jump Your Steps}$: Optimizing Sampling Schedule of Discrete Diffusion Models 2024 Yong-Hyun Park
Chieh-Hsin Lai
Satoshi Hayakawa
Yuhta Takida
Yuki Mitsufuji

Works That Cite This (0)

Action Title Year Authors

Works Cited by This (0)

Action Title Year Authors