Precise asymptotics for phase retrieval and compressed sensing with random generative priors
NeurIPS 2019 Workshop on Solving Inverse Problems with Deep Networks, 2019•openreview.net
We consider the problem of compressed sensing and of (real-valued) phase retrieval with
random measurement matrix. We analyse sharp asymptotics of the information-theoretically
optimal performance and that of the best known polynomial algorithms under a generative
prior consisting of a single layer neural network with a random weight matrix. We compare
the performance to sparse separable priors and conclude that generative priors might be
advantageous in terms of algorithmic performance. In particular, while sparsity does not …
random measurement matrix. We analyse sharp asymptotics of the information-theoretically
optimal performance and that of the best known polynomial algorithms under a generative
prior consisting of a single layer neural network with a random weight matrix. We compare
the performance to sparse separable priors and conclude that generative priors might be
advantageous in terms of algorithmic performance. In particular, while sparsity does not …
We consider the problem of compressed sensing and of (real-valued) phase retrieval with random measurement matrix. We analyse sharp asymptotics of the information-theoretically optimal performance and that of the best known polynomial algorithms under a generative prior consisting of a single layer neural network with a random weight matrix. We compare the performance to sparse separable priors and conclude that generative priors might be advantageous in terms of algorithmic performance. In particular, while sparsity does not allow to perform compressive phase retrieval efficiently close to its information-theoretic limit, it is found that under the random generative prior compressed phase retrieval becomes tractable.
openreview.net
以上显示的是最相近的搜索结果。 查看全部搜索结果