尹志 (2022-04-28 22:10):
#paper https://doi.org/10.48550/arXiv.1503.03585 Deep Unsupervised Learning using Nonequilibrium Thermodynamics ICML (2015). 这是一篇还没完全看懂的论文,但是非常有意思。说起这篇文章的扩散模型大家一不定熟悉,但是提到最近大火的openai的工作dall-e 2,大家可能会更熟悉一点。对,Dall-E 2最早的启发就是这篇文章。本文受非平衡热力学的启发,设计了一个称之为扩散模型(diffusion model)的生成模型。我们知道,在机器学习中,对一堆数据的分布进行估计是一个极具挑战的事情。特别是要兼顾模型的灵活性(flexible)和过程的可解性(tractable)。如果把建模隐变量z到观测量x的映射作为任务,那么扩散模型的想法是, 假设整个映射是一个马尔科夫链(MC),然后数据的初始状态是由一步步不断添加高斯噪声,最终获得某种最终形态,那么反过来,可以将去噪的过程看做是生成的过程。我们针对这个MC过程进行训练,那么逆过程则可以作为生成模型生成符合分布的数据。是的,很像VAE。考虑到这类生成模型通过不断的改进,已经达到Dall-E 2的效果,值得我们深入理解背后的机制,以及是否可以在数据合成上产生更好的效果。
Deep Unsupervised Learning using Nonequilibrium Thermodynamics
翻译
Abstract:
A central problem in machine learning involves modeling complex data-sets using highly flexible families of probability distributions in which learning, sampling, inference, and evaluation are still analytically or computationally tractable. Here, we develop an approach that simultaneously achieves both flexibility and tractability. The essential idea, inspired by non-equilibrium statistical physics, is to systematically and slowly destroy structure in a data distribution through an iterative forward diffusion process. We then learn a reverse diffusion process that restores structure in data, yielding a highly flexible and tractable generative model of the data. This approach allows us to rapidly learn, sample from, and evaluate probabilities in deep generative models with thousands of layers or time steps, as well as to compute conditional and posterior probabilities under the learned model. We additionally release an open source reference implementation of the algorithm.
翻译
回到顶部