Entropy and mutual information in models of deep neural networks
We examine a class of stochastic deep learning models with a tractable method to compute
information-theoretic quantities. Our contributions are three-fold:(i) We show how entropies …
information-theoretic quantities. Our contributions are three-fold:(i) We show how entropies …
Rigorous dynamics of expectation-propagation-based signal recovery from unitarily invariant measurements
K Takeuchi - IEEE Transactions on Information Theory, 2019 - ieeexplore.ieee.org
Signal recovery from unitarily invariant measurements is investigated in this paper. A
message-passing algorithm is formulated on the basis of expectation propagation (EP). A …
message-passing algorithm is formulated on the basis of expectation propagation (EP). A …
The adaptive interpolation method: a simple scheme to prove replica formulas in Bayesian inference
In recent years important progress has been achieved towards proving the validity of the
replica predictions for the (asymptotic) mutual information (or “free energy”) in Bayesian …
replica predictions for the (asymptotic) mutual information (or “free energy”) in Bayesian …
The adaptive interpolation method for proving replica formulas. Applications to the Curie–Weiss and Wigner spike models
In this contribution we give a pedagogic introduction to the newly introduced adaptive
interpolation method to prove in a simple and unified way replica formulas for Bayesian …
interpolation method to prove in a simple and unified way replica formulas for Bayesian …
Phase retrieval in high dimensions: Statistical and computational phase transitions
We consider the phase retrieval problem of reconstructing a $ n $-dimensional real or
complex signal $\mathbf {X}^\star $ from $ m $(possibly noisy) observations $ Y_\mu=|\sum …
complex signal $\mathbf {X}^\star $ from $ m $(possibly noisy) observations $ Y_\mu=|\sum …
On double-descent in uncertainty quantification in overparametrized models
Uncertainty quantification is a central challenge in reliable and trustworthy machine
learning. Naive measures such as last-layer scores are well-known to yield overconfident …
learning. Naive measures such as last-layer scores are well-known to yield overconfident …
Memory amp
L Liu, S Huang, BM Kurkoski - IEEE Transactions on …, 2022 - ieeexplore.ieee.org
Approximate message passing (AMP) is a low-cost iterative parameter-estimation technique
for certain high-dimensional linear systems with non-Gaussian distributions. AMP only …
for certain high-dimensional linear systems with non-Gaussian distributions. AMP only …
Bayes-optimal convolutional AMP
K Takeuchi - IEEE Transactions on Information Theory, 2021 - ieeexplore.ieee.org
This paper proposes Bayes-optimal convolutional approximate message-passing (CAMP)
for signal recovery in compressed sensing. CAMP uses the same low-complexity matched …
for signal recovery in compressed sensing. CAMP uses the same low-complexity matched …
The price of ignorance: how much does it cost to forget noise structure in low-rank matrix estimation?
We consider the problem of estimating a rank-$1 $ signal corrupted by structured rotationally
invariant noise, and address the following question:\emph {how well do inference algorithms …
invariant noise, and address the following question:\emph {how well do inference algorithms …
On the convergence of orthogonal/vector AMP: Long-memory message-passing strategy
K Takeuchi - IEEE Transactions on Information Theory, 2022 - ieeexplore.ieee.org
Orthogonal/vector approximate message-passing (AMP) is a powerful message-passing
(MP) algorithm for signal reconstruction in compressed sensing. This paper proves the …
(MP) algorithm for signal reconstruction in compressed sensing. This paper proves the …