Holistic evaluation of language models P Liang, R Bommasani, T Lee, D Tsipras, D Soylu, M Yasunaga, Y Zhang, ... arXiv preprint arXiv:2211.09110, 2022 | 795 | 2022 |
Two are Better than One: Joint Entity and Relation Extraction with Table-Sequence Encoders J Wang, W Lu Empirical Methods in Natural Language Processing, 1706--1721, 2020 | 226 | 2020 |
Pyramid: A Layered Model for Nested Named Entity Recognition J Wang, L Shou, K Chen, G Chen Association for Computational Linguistics, 5918–5928, 2020 | 151 | 2020 |
Deja vu: Contextual sparsity for efficient llms at inference time Z Liu, J Wang, T Dao, T Zhou, B Yuan, Z Song, A Shrivastava, C Zhang, ... | 123 | 2023 |
Continual Federated Learning Based on Knowledge Distillation. Y Ma, Z Xie, J Wang, K Chen, L Shou IJCAI, 2182-2188, 2022 | 36 | 2022 |
Draft & verify: Lossless large language model acceleration via self-speculative decoding J Zhang, J Wang, H Li, L Shou, K Chen, G Chen, S Mehrotra arXiv preprint arXiv:2309.08168, 2023 | 27 | 2023 |
Skill-it! a data-driven skills framework for understanding and training language models M Chen, N Roberts, K Bhatia, J Wang, C Zhang, F Sala, C Ré Advances in Neural Information Processing Systems 36, 2024 | 22 | 2024 |
Skipbert: Efficient inference with shallow layer skipping J Wang, K Chen, G Chen, L Shou, J McAuley Proceedings of the 60th Annual Meeting of the Association for Computational …, 2022 | 20 | 2022 |
CocktailSGD: Fine-tuning Foundation Models over 500Mbps Networks J Wang, Y Lu, B Yuan, B Chen, P Liang, C De Sa, C Re, C Zhang | 19 | 2023 |
Compress, then prompt: Improving accuracy-efficiency trade-off of llm inference with transferable prompt Z Xu, Z Liu, B Chen, Y Tang, J Wang, K Zhou, X Hu, A Shrivastava arXiv preprint arXiv:2305.11186, 2023 | 18 | 2023 |
Fine-tuning language models over slow networks using activation quantization with guarantees J Wang, B Yuan, L Rimanic, Y He, T Dao, B Chen, C Ré, C Zhang Advances in Neural Information Processing Systems 35, 19215-19230, 2022 | 13* | 2022 |
Mixture-of-Agents Enhances Large Language Model Capabilities J Wang, J Wang, B Athiwaratkun, C Zhang, J Zou arXiv preprint arXiv:2406.04692, 2024 | 4 | 2024 |
Effective Slot Filling via Weakly-Supervised Dual-Model Learning J Wang, K Chen, L Shou, S Wu, G Chen AAAI Conference on Artificial Intelligence 35 (16), 13952-13960, 2021 | 4 | 2021 |
Effective Continual Learning for Text Classification with Lightweight Snapshots J Wang, D Dong, L Shou, K Chen, G Chen AAAI Conference on Artificial Intelligence 37, 10122-10130, 2023 | 2 | 2023 |
Semi-supervised few-shot learning for dual question-answer extraction J Wang, K Chen, L Shou, S Wu, S Mehrotra arXiv preprint arXiv:1904.03898, 2019 | 2 | 2019 |
Learning Label-Adaptive Representation for Large-Scale Multi-Label Text Classification C Peng, H Wang, J Wang, L Shou, K Chen, G Chen, C Yao IEEE/ACM Transactions on Audio, Speech, and Language Processing, 2024 | | 2024 |
SMILE: A Cost-Effective System for Serving Massive Pretrained Language Models in The Cloud J Wang, K Chen, L Shou, D Jiang, G Chen Companion of the 2023 International Conference on Management of Data, 135-138, 2023 | | 2023 |