[PDF][PDF] Efficient large language models: A survey
Abstract Large Language Models (LLMs) have demonstrated remarkable capabilities in
important tasks such as natural language understanding, language generation, and …
important tasks such as natural language understanding, language generation, and …
Beyond efficiency: A systematic survey of resource-efficient large language models
The burgeoning field of Large Language Models (LLMs), exemplified by sophisticated
models like OpenAI's ChatGPT, represents a significant advancement in artificial …
models like OpenAI's ChatGPT, represents a significant advancement in artificial …
A survey on efficient training of transformers
Recent advances in Transformers have come with a huge requirement on computing
resources, highlighting the importance of developing efficient training techniques to make …
resources, highlighting the importance of developing efficient training techniques to make …
A survey of resource-efficient llm and multimodal foundation models
Large foundation models, including large language models (LLMs), vision transformers
(ViTs), diffusion, and LLM-based multimodal models, are revolutionizing the entire machine …
(ViTs), diffusion, and LLM-based multimodal models, are revolutionizing the entire machine …
Memory efficient optimizers with 4-bit states
Optimizer states are a major source of memory consumption for training neural networks,
limiting the maximum trainable model within given memory budget. Compressing the …
limiting the maximum trainable model within given memory budget. Compressing the …
Winner-take-all column row sampling for memory efficient adaptation of language model
As the model size grows rapidly, fine-tuning the large pre-trained language model has
become increasingly difficult due to its extensive memory usage. Previous works usually …
become increasingly difficult due to its extensive memory usage. Previous works usually …
Tinytrain: Deep neural network training at the extreme edge
On-device training is essential for user personalisation and privacy. With the pervasiveness
of IoT devices and microcontroller units (MCU), this task becomes more challenging due to …
of IoT devices and microcontroller units (MCU), this task becomes more challenging due to …
Tinykg: Memory-efficient training framework for knowledge graph neural recommender systems
There has been an explosion of interest in designing various Knowledge Graph Neural
Networks (KGNNs), which achieve state-of-the-art performance and provide great …
Networks (KGNNs), which achieve state-of-the-art performance and provide great …
TANGO: re-thinking quantization for graph neural network training on GPUs
Graph learning is becoming increasingly popular due to its superior performance in tackling
many grand challenges. While quantization is widely used to accelerate Graph Neural …
many grand challenges. While quantization is widely used to accelerate Graph Neural …
DIVISION: memory efficient training via dual activation precision
Activation compressed training provides a solution towards reducing the memory cost of
training deep neural networks (DNNs). However, state-of-the-art work combines a search of …
training deep neural networks (DNNs). However, state-of-the-art work combines a search of …