关注
Simiao Zuo
Simiao Zuo
在 gatech.edu 的电子邮件经过验证 - 首页
标题
引用次数
引用次数
年份
Transformer hawkes process
S Zuo, H Jiang, Z Li, T Zhao, H Zha
International Conference on Machine Learning, 11692-11702, 2020
2792020
Fine-Tuning Pre-trained Language Model with Weak Supervision: A Contrastive-Regularized Self-Training Approach
Y Yu, S Zuo, H Jiang, W Ren, T Zhao, C Zhang
arXiv preprint arXiv:2010.07835, 2020
1132020
Taming Sparsely Activated Transformer with Stochastic Experts
S Zuo, X Liu, J Jiao, YJ Kim, H Hassan, R Zhang, T Zhao, J Gao
arXiv preprint arXiv:2110.04260, 2021
832021
PLATON: Pruning Large Transformer Models with Upper Confidence Bound of Weight Importance
Q Zhang, S Zuo, C Liang, A Bukharin, P He, W Chen, T Zhao
International Conference on Machine Learning, 26809-26823, 2022
552022
Super Tickets in Pre-Trained Language Models: From Model Compression to Improving Generalization
C Liang, S Zuo, M Chen, H Jiang, X Liu, P He, T Zhao, W Chen
arXiv preprint arXiv:2105.12002, 2021
482021
Less is More: Task-aware Layer-wise Distillation for Language Model Compression
C Liang, S Zuo, Q Zhang, P He, W Chen, T Zhao
arXiv preprint arXiv:2210.01351, 2022
352022
MoEBERT: from BERT to Mixture-of-Experts via Importance-Guided Adaptation
S Zuo, Q Zhang, C Liang, P He, T Zhao, W Chen
arXiv preprint arXiv:2204.07675, 2022
292022
Efficient Long Sequence Modeling via State Space Augmented Transformer
S Zuo, X Liu, J Jiao, D Charles, E Manavoglu, T Zhao, J Gao
arXiv preprint arXiv:2212.08136, 2022
272022
A Hypergradient Approach to Robust Regression without Correspondence
Y Xie, Y Mao, S Zuo, H Xu, X Ye, T Zhao, H Zha
arXiv preprint arXiv:2012.00123, 2020
152020
No Parameters Left Behind: Sensitivity Guided Adaptive Learning Rate for Training Large Transformer Models
C Liang, H Jiang, S Zuo, P He, X Liu, J Gao, W Chen, T Zhao
arXiv preprint arXiv:2202.02664, 2022
142022
Self-Training with Differentiable Teacher
S Zuo, Y Yu, C Liang, H Jiang, S Er, C Zhang, T Zhao, H Zha
arXiv preprint arXiv:2109.07049, 2021
142021
Adversarial training as stackelberg game: An unrolled optimization approach
S Zuo, C Liang, H Jiang, X Liu, P He, J Gao, W Chen, T Zhao
arXiv preprint arXiv:2104.04886, 2021
14*2021
Tensor maps for synchronizing heterogeneous shape collections
Q Huang, Z Liang, H Wang, S Zuo, C Bajaj
ACM Transactions on Graphics (TOG) 38 (4), 1-18, 2019
122019
Adversarially regularized policy learning guided by trajectory optimization
Z Zhao, S Zuo, T Zhao, Y Zhao
Learning for Dynamics and Control Conference, 844-857, 2022
112022
Shaohui Xi, Bing Yin, Chao Zhang, and Tuo Zhao. 2022. Context-Aware Query Rewriting for Improving Users’ Search Experience on E-commerce Websites
S Zuo, Q Yin, H Jiang
arXiv preprint arXiv:2209.07584, 2022
62022
Robust Multi-Agent Reinforcement Learning via Adversarial Regularization: Theoretical Foundation and Stable Algorithms
A Bukharin, Y Li, Y Yu, Q Zhang, Z Chen, S Zuo, C Zhang, S Zhang, ...
arXiv preprint arXiv:2310.10810, 2023
52023
ARCH: Efficient Adversarial Regularized Training with Caching
S Zuo, C Liang, H Jiang, P He, X Liu, J Gao, W Chen, T Zhao
arXiv preprint arXiv:2109.07048, 2021
32021
SMURF-THP: Score Matching-based UnceRtainty quantiFication for Transformer Hawkes Process
Z Li, Y Xu, S Zuo, H Jiang, C Zhang, T Zhao, H Zha
22023
DiP-GNN: Discriminative Pre-Training of Graph Neural Networks
S Zuo, H Jiang, Q Yin, X Tang, B Yin, T Zhao
arXiv preprint arXiv:2209.07499, 2022
22022
Differentially Private Estimation of Hawkes Process
S Zuo, T Liu, T Zhao, H Zha
arXiv preprint arXiv:2209.07303, 2022
22022
系统目前无法执行此操作,请稍后再试。
文章 1–20