Towards building the federatedGPT: Federated instruction tuning
While" instruction-tuned" generative large language models (LLMs) have demonstrated an
impressive ability to generalize to new tasks, the training phases heavily rely on large …
impressive ability to generalize to new tasks, the training phases heavily rely on large …
Fed-cbs: A heterogeneity-aware client sampling mechanism for federated learning via class-imbalance reduction
Due to the often limited communication bandwidth of edge devices, most existing federated
learning (FL) methods randomly select only a subset of devices to participate in training at …
learning (FL) methods randomly select only a subset of devices to participate in training at …
Dpsur: Accelerating differentially private stochastic gradient descent using selective update and release
Machine learning models are known to memorize private data to reduce their training loss,
which can be inadvertently exploited by privacy attacks such as model inversion and …
which can be inadvertently exploited by privacy attacks such as model inversion and …
Training Bayesian Neural Networks with Sparse Subspace Variational Inference
Bayesian neural networks (BNNs) offer uncertainty quantification but come with the
downside of substantially increased training and inference costs. Sparse BNNs have been …
downside of substantially increased training and inference costs. Sparse BNNs have been …
Particle-based online bayesian sampling
Online optimization has gained increasing interest due to its capability of tracking real-world
streaming data. Although online optimization methods have been widely studied in the …
streaming data. Although online optimization methods have been widely studied in the …
Bayesian Multi-Task Transfer Learning for Soft Prompt Tuning
Prompt tuning, in which prompts are optimized to adapt large-scale pre-trained language
models to downstream tasks instead of fine-tuning the full model parameters, has been …
models to downstream tasks instead of fine-tuning the full model parameters, has been …
Variance reduction and quasi-Newton for particle-based variational inference
Abstract Particle-based Variational Inference methods (ParVIs), like Stein Variational
Gradient Descent, are nonparametric variational inference methods that optimize a set of …
Gradient Descent, are nonparametric variational inference methods that optimize a set of …
Dpvi: A dynamic-weight particle-based variational inference framework
The recently developed Particle-based Variational Inference (ParVI) methods drive the
empirical distribution of a set of\emph {fixed-weight} particles towards a given target …
empirical distribution of a set of\emph {fixed-weight} particles towards a given target …
SLED: Self Logits Evolution Decoding for Improving Factuality in Large Language Models
Large language models (LLMs) have demonstrated remarkable capabilities, but their
outputs can sometimes be unreliable or factually incorrect. To address this, we introduce …
outputs can sometimes be unreliable or factually incorrect. To address this, we introduce …
Towards uncertainty and efficiency in reinforcement learning
R Zhang - 2021 - search.proquest.com
Deep reinforcement learning (RL) has received great success in playing video games and
strategic board games, where a simulator is well-defined, and massive samples are …
strategic board games, where a simulator is well-defined, and massive samples are …