Randomized algorithms for computation of Tucker decomposition and higher order SVD (HOSVD)
Big data analysis has become a crucial part of new emerging technologies such as the
internet of things, cyber-physical analysis, deep learning, anomaly detection, etc. Among …
internet of things, cyber-physical analysis, deep learning, anomaly detection, etc. Among …
A literature survey of matrix methods for data science
M Stoll - GAMM‐Mitteilungen, 2020 - Wiley Online Library
Efficient numerical linear algebra is a core ingredient in many applications across almost all
scientific and industrial disciplines. With this survey we want to illustrate that numerical linear …
scientific and industrial disciplines. With this survey we want to illustrate that numerical linear …
Towards compact neural networks via end-to-end training: A Bayesian tensor approach with automatic rank determination
Post-training model compression can reduce the inference costs of deep neural networks,
but uncompressed training still consumes enormous hardware resources and energy. To …
but uncompressed training still consumes enormous hardware resources and energy. To …
Training acceleration of low-rank decomposed networks using sequential freezing and rank quantization
H Hajimolahoseini, W Ahmed, Y Liu - arXiv preprint arXiv:2309.03824, 2023 - arxiv.org
Low Rank Decomposition (LRD) is a model compression technique applied to the weight
tensors of deep learning models in order to reduce the number of trainable parameters and …
tensors of deep learning models in order to reduce the number of trainable parameters and …
[PDF][PDF] Strategies for applying low rank decomposition to transformer-based models
H Hajimolahoseini, W Ahmed… - 36th Conference …, 2022 - neurips2022-enlsp.github.io
Low rank decomposition decomposes each fully-connected layer of the transformer modules
into two smaller layers using Singular Value Decomposition. The state-of-the-art techniques …
into two smaller layers using Singular Value Decomposition. The state-of-the-art techniques …
Reduced-order modeling of deep neural networks
We introduce a new method for speeding up the inference of deep neural networks. It is
somewhat inspired by the reduced-order modeling techniques for dynamical systems. The …
somewhat inspired by the reduced-order modeling techniques for dynamical systems. The …
Edge AI–A Promising Technology
Summary Edge Artificial Intelligence (Edge AI) has become the buzzword for every industry
organization. Edge intelligence utilizes edge computing to access and analyze the data from …
organization. Edge intelligence utilizes edge computing to access and analyze the data from …
Octave deep compression: In-parallel pruning-quantization on different frequencies
Though deep neural networks achieve great accuracy in visual recognition tasks, they
contain millions of weights and thus require a large space to be stored. This presents a …
contain millions of weights and thus require a large space to be stored. This presents a …
[图书][B] Compressed Training for Uncertainty-Aware Compact Neural Networks
CP Hawkins - 2022 - search.proquest.com
The rising computational and memory demands of machine learning models, particularly in
resource-constrained edge-device settings, motivate us to develop compressed models that …
resource-constrained edge-device settings, motivate us to develop compressed models that …
[PDF][PDF] Compressing neural networks through CP-decomposition
T Rudkiewicz - 2023 - perso.crans.org
Today neural networks are state of the art for numerous task including image classification.
The best neural networks use an enormous amount of parameters. Some works try to reduce …
The best neural networks use an enormous amount of parameters. Some works try to reduce …