Towards building the federatedGPT: Federated instruction tuning

J Zhang, S Vahidian, M Kuo, C Li… - ICASSP 2024-2024 …, 2024 - ieeexplore.ieee.org
While" instruction-tuned" generative large language models (LLMs) have demonstrated an
impressive ability to generalize to new tasks, the training phases heavily rely on large …

Fed-cbs: A heterogeneity-aware client sampling mechanism for federated learning via class-imbalance reduction

J Zhang, A Li, M Tang, J Sun, X Chen… - International …, 2023 - proceedings.mlr.press
Due to the often limited communication bandwidth of edge devices, most existing federated
learning (FL) methods randomly select only a subset of devices to participate in training at …

Dpsur: Accelerating differentially private stochastic gradient descent using selective update and release

J Fu, Q Ye, H Hu, Z Chen, L Wang, K Wang… - arXiv preprint arXiv …, 2023 - arxiv.org
Machine learning models are known to memorize private data to reduce their training loss,
which can be inadvertently exploited by privacy attacks such as model inversion and …

Training Bayesian Neural Networks with Sparse Subspace Variational Inference

J Li, Z Miao, Q Qiu, R Zhang - arXiv preprint arXiv:2402.11025, 2024 - arxiv.org
Bayesian neural networks (BNNs) offer uncertainty quantification but come with the
downside of substantially increased training and inference costs. Sparse BNNs have been …

Particle-based online bayesian sampling

Y Yang, C Liu, Z Zhang - arXiv preprint arXiv:2302.14796, 2023 - arxiv.org
Online optimization has gained increasing interest due to its capability of tracking real-world
streaming data. Although online optimization methods have been widely studied in the …

Bayesian Multi-Task Transfer Learning for Soft Prompt Tuning

H Lee, M Jeong, SY Yun, KE Kim - arXiv preprint arXiv:2402.08594, 2024 - arxiv.org
Prompt tuning, in which prompts are optimized to adapt large-scale pre-trained language
models to downstream tasks instead of fine-tuning the full model parameters, has been …

Variance reduction and quasi-Newton for particle-based variational inference

M Zhu, C Liu, J Zhu - International Conference on Machine …, 2020 - proceedings.mlr.press
Abstract Particle-based Variational Inference methods (ParVIs), like Stein Variational
Gradient Descent, are nonparametric variational inference methods that optimize a set of …

Dpvi: A dynamic-weight particle-based variational inference framework

C Zhang, Z Li, H Qian, X Du - arXiv preprint arXiv:2112.00945, 2021 - arxiv.org
The recently developed Particle-based Variational Inference (ParVI) methods drive the
empirical distribution of a set of\emph {fixed-weight} particles towards a given target …

SLED: Self Logits Evolution Decoding for Improving Factuality in Large Language Models

J Zhang, DC Juan, C Rashtchian, CS Ferng… - arXiv preprint arXiv …, 2024 - arxiv.org
Large language models (LLMs) have demonstrated remarkable capabilities, but their
outputs can sometimes be unreliable or factually incorrect. To address this, we introduce …

Towards uncertainty and efficiency in reinforcement learning

R Zhang - 2021 - search.proquest.com
Deep reinforcement learning (RL) has received great success in playing video games and
strategic board games, where a simulator is well-defined, and massive samples are …