PAC-Bayes compression bounds so tight that they can explain generalization
While there has been progress in developing non-vacuous generalization bounds for deep
neural networks, these bounds tend to be uninformative about why deep learning works. In …
neural networks, these bounds tend to be uninformative about why deep learning works. In …
Reconstruction for powerful graph representations
Graph neural networks (GNNs) have limited expressive power, failing to represent many
graph classes correctly. While more expressive graph representation learning (GRL) …
graph classes correctly. While more expressive graph representation learning (GRL) …
Improving self-supervised learning by characterizing idealized representations
Despite the empirical successes of self-supervised learning (SSL) methods, it is unclear
what characteristics of their representations lead to high downstream accuracies. In this …
what characteristics of their representations lead to high downstream accuracies. In this …
Probabilistic symmetries and invariant neural networks
B Bloem-Reddy, Y Whye - Journal of Machine Learning Research, 2020 - jmlr.org
Treating neural network inputs and outputs as random variables, we characterize the
structure of neural networks that can be used to model data that are invariant or equivariant …
structure of neural networks that can be used to model data that are invariant or equivariant …
Lossy compression for lossless prediction
Most data is automatically collected and only ever" seen" by algorithms. Yet, data
compressors preserve perceptual fidelity rather than just the information needed by …
compressors preserve perceptual fidelity rather than just the information needed by …
Out-of-domain robustness via targeted augmentations
Abstract Models trained on one set of domains often suffer performance drops on unseen
domains, eg, when wildlife monitoring models are deployed in new camera locations. In this …
domains, eg, when wildlife monitoring models are deployed in new camera locations. In this …
Approximately equivariant graph networks
Graph neural networks (GNNs) are commonly described as being permutation equivariant
with respect to node relabeling in the graph. This symmetry of GNNs is often compared to …
with respect to node relabeling in the graph. This symmetry of GNNs is often compared to …
Causally motivated shortcut removal using auxiliary labels
Shortcut learning, in which models make use of easy-to-represent but unstable associations,
is a major failure mode for robust machine learning. We study a flexible, causally-motivated …
is a major failure mode for robust machine learning. We study a flexible, causally-motivated …
Approximation-generalization trade-offs under (approximate) group equivariance
M Petrache, S Trivedi - Advances in Neural Information …, 2023 - proceedings.neurips.cc
The explicit incorporation of task-specific inductive biases through symmetry has emerged
as a general design precept in the development of high-performance machine learning …
as a general design precept in the development of high-performance machine learning …
HYTREL: Hypergraph-enhanced tabular data representation learning
Abstract Language models pretrained on large collections of tabular data have
demonstrated their effectiveness in several downstream tasks. However, many of these …
demonstrated their effectiveness in several downstream tasks. However, many of these …