AutoLoRA: AutoGuidance Meets Low-Rank Adaptation for Diffusion Models

A Kasymov, M Sendera, M Stypułkowski… - arXiv preprint arXiv …, 2024 - arxiv.org
Low-rank adaptation (LoRA) is a fine-tuning technique that can be applied to conditional
generative diffusion models. LoRA utilizes a small number of context examples to adapt the …

Amortizing intractable inference in diffusion models for vision, language, and control

S Venkatraman, M Jain, L Scimeca, M Kim… - arXiv preprint arXiv …, 2024 - arxiv.org
Diffusion models have emerged as effective distribution estimators in vision, language, and
reinforcement learning, but their use as priors in downstream tasks poses an intractable …

Flow of Reasoning: Efficient Training of LLM Policy with Divergent Thinking

F Yu, L Jiang, H Kang, S Hao, L Qin - arXiv preprint arXiv:2406.05673, 2024 - arxiv.org
Divergent thinking, the cognitive process of generating diverse solutions, is a hallmark of
human creativity and problem-solving. For machines, sampling diverse solution trajectories …

Steering Masked Discrete Diffusion Models via Discrete Denoising Posterior Prediction

J Rector-Brooks, M Hasan, Z Peng, Z Quinn… - arXiv preprint arXiv …, 2024 - arxiv.org
Generative modeling of discrete data underlies important applications spanning text-based
agents like ChatGPT to the design of the very building blocks of life in protein sequences …

Understanding and mitigating difficulties in posterior predictive evaluation

A Agrawal, J Domke - arXiv preprint arXiv:2405.19747, 2024 - arxiv.org
Predictive posterior densities (PPDs) are of interest in approximate Bayesian inference.
Typically, these are estimated by simple Monte Carlo (MC) averages using samples from the …

Sequential Controlled Langevin Diffusions

J Chen, L Richter, J Berner, D Blessing… - arXiv preprint arXiv …, 2024 - arxiv.org
An effective approach for sampling from unnormalized densities is based on the idea of
gradually transporting samples from an easy prior to the complicated target distribution. Two …

Pessimistic Backward Policy for GFlowNets

H Jang, Y Jang, M Kim, J Park, S Ahn - arXiv preprint arXiv:2405.16012, 2024 - arxiv.org
This paper studies Generative Flow Networks (GFlowNets), which learn to sample objects
proportionally to a given reward function through the trajectory of state transitions. In this …

Beyond ELBOs: A Large-Scale Evaluation of Variational Methods for Sampling

D Blessing, X Jia, J Esslinger, F Vargas… - arXiv preprint arXiv …, 2024 - arxiv.org
Monte Carlo methods, Variational Inference, and their combinations play a pivotal role in
sampling from intractable probability distributions. However, current studies lack a unified …

Streaming Bayes GFlowNets

T da Silva, DA de Souza, D Mesquita - arXiv preprint arXiv:2411.05899, 2024 - arxiv.org
Bayes' rule naturally allows for inference refinement in a streaming fashion, without the need
to recompute posteriors from scratch whenever new data arrives. In principle, Bayesian …

Iterated Energy-based Flow Matching for Sampling from Boltzmann Densities

D Woo, S Ahn - arXiv preprint arXiv:2408.16249, 2024 - arxiv.org
In this work, we consider the problem of training a generator from evaluations of energy
functions or unnormalized densities. This is a fundamental problem in probabilistic …