Model compression and hardware acceleration for neural networks: A comprehensive survey

L Deng, G Li, S Han, L Shi, Y Xie - Proceedings of the IEEE, 2020 - ieeexplore.ieee.org
Domain-specific hardware is becoming a promising topic in the backdrop of improvement
slow down for general-purpose processors due to the foreseeable end of Moore's Law …

A survey on deep learning for multimodal data fusion

J Gao, P Li, Z Chen, J Zhang - Neural Computation, 2020 - direct.mit.edu
With the wide deployments of heterogeneous networks, huge amounts of data with
characteristics of high volume, high variety, high velocity, and high veracity are generated …

Mamba: Linear-time sequence modeling with selective state spaces

A Gu, T Dao - arXiv preprint arXiv:2312.00752, 2023 - arxiv.org
Foundation models, now powering most of the exciting applications in deep learning, are
almost universally based on the Transformer architecture and its core attention module …

Rwkv: Reinventing rnns for the transformer era

B Peng, E Alcaide, Q Anthony, A Albalak… - arXiv preprint arXiv …, 2023 - arxiv.org
Transformers have revolutionized almost all natural language processing (NLP) tasks but
suffer from memory and computational complexity that scales quadratically with sequence …

[HTML][HTML] Artificial intelligence-enabled detection and assessment of Parkinson's disease using nocturnal breathing signals

Y Yang, Y Yuan, G Zhang, H Wang, YC Chen, Y Liu… - Nature medicine, 2022 - nature.com
There are currently no effective biomarkers for diagnosing Parkinson's disease (PD) or
tracking its progression. Here, we developed an artificial intelligence (AI) model to detect PD …

Simplified state space layers for sequence modeling

JTH Smith, A Warrington, SW Linderman - arXiv preprint arXiv:2208.04933, 2022 - arxiv.org
Models using structured state space sequence (S4) layers have achieved state-of-the-art
performance on long-range sequence modeling tasks. An S4 layer combines linear state …

Combining recurrent, convolutional, and continuous-time models with linear state space layers

A Gu, I Johnson, K Goel, K Saab… - Advances in neural …, 2021 - proceedings.neurips.cc
Recurrent neural networks (RNNs), temporal convolutions, and neural differential equations
(NDEs) are popular families of deep learning models for time-series data, each with unique …

Underspecification presents challenges for credibility in modern machine learning

A D'Amour, K Heller, D Moldovan, B Adlam… - Journal of Machine …, 2022 - jmlr.org
Machine learning (ML) systems often exhibit unexpectedly poor behavior when they are
deployed in real-world domains. We identify underspecification in ML pipelines as a key …

Hierarchically gated recurrent neural network for sequence modeling

Z Qin, S Yang, Y Zhong - Advances in Neural Information …, 2024 - proceedings.neurips.cc
Transformers have surpassed RNNs in popularity due to their superior abilities in parallel
training and long-term dependency modeling. Recently, there has been a renewed interest …

Delving into deep imbalanced regression

Y Yang, K Zha, Y Chen, H Wang… - … conference on machine …, 2021 - proceedings.mlr.press
Real-world data often exhibit imbalanced distributions, where certain target values have
significantly fewer observations. Existing techniques for dealing with imbalanced data focus …