An information-theoretic perspective on variance-invariance-covariance regularization

R Shwartz-Ziv, R Balestriero, K Kawaguchi… - arXiv preprint arXiv …, 2023 - arxiv.org
In this paper, we provide an information-theoretic perspective on Variance-Invariance-
Covariance Regularization (VICReg) for self-supervised learning. To do so, we first …

An information theory perspective on variance-invariance-covariance regularization

R Shwartz-Ziv, R Balestriero… - Advances in …, 2023 - proceedings.neurips.cc
Abstract Variance-Invariance-Covariance Regularization (VICReg) is a self-supervised
learning (SSL) method that has shown promising results on a variety of tasks. However, the …

martFL: Enabling Utility-Driven Data Marketplace with a Robust and Verifiable Federated Learning Architecture

Q Li, Z Liu, Q Li, K Xu - Proceedings of the 2023 ACM SIGSAC …, 2023 - dl.acm.org
The development of machine learning models requires a large amount of training data. Data
marketplace is a critical platform to trade high-quality and private-domain data that is not …

Fine-grained data distribution alignment for post-training quantization

Y Zhong, M Lin, M Chen, K Li, Y Shen, F Chao… - … on Computer Vision, 2022 - Springer
While post-training quantization receives popularity mostly due to its evasion in accessing
the original complete training dataset, its poor performance also stems from scarce images …

Vq4dit: Efficient post-training vector quantization for diffusion transformers

J Deng, S Li, Z Wang, H Gu, K Xu, K Huang - arXiv preprint arXiv …, 2024 - arxiv.org
The Diffusion Transformers Models (DiTs) have transitioned the network architecture from
traditional UNets to transformers, demonstrating exceptional capabilities in image …

Sub-8-bit quantization for on-device speech recognition: A regularization-free approach

K Zhen, M Radfar, H Nguyen, GP Strimel… - 2022 IEEE Spoken …, 2023 - ieeexplore.ieee.org
For on-device automatic speech recognition (ASR), quantization aware training (QAT) is
ubiquitous to achieve the trade-off between model predictive performance and efficiency …

Yono: Modeling multiple heterogeneous neural networks on microcontrollers

YD Kwon, J Chauhan, C Mascolo - 2022 21st ACM/IEEE …, 2022 - ieeexplore.ieee.org
Internet of Things (IoT) systems provide large amounts of data on all aspects of human
behavior. Machine learning techniques, especially deep neural networks (DNN), have …

Enabling on-device smartphone GPU based training: Lessons learned

A Das, YD Kwon, J Chauhan… - 2022 IEEE International …, 2022 - ieeexplore.ieee.org
Deep Learning (DL) has shown impressive performance in many mobile applications. Most
existing works have focused on reducing the computational and resource overheads of …

A noise-driven heterogeneous stochastic computing multiplier for heuristic precision improvement in energy-efficient dnns

J Wang, H Chen, D Wang, K Mei… - IEEE Transactions on …, 2022 - ieeexplore.ieee.org
Stochastic computing (SC) has become a promising approximate computing solution by its
negligible resource occupancy and ultralow energy consumption. As a potential …

Gptvq: The blessing of dimensionality for llm quantization

M van Baalen, A Kuzmin, M Nagel, P Couperus… - arXiv preprint arXiv …, 2024 - arxiv.org
In this work we show that the size versus accuracy trade-off of neural network quantization
can be significantly improved by increasing the quantization dimensionality. We propose the …