Improved parallel matrix multiplication using Strassen and Urdhvatiryagbhyam method

YRA Bessant, JG Jency, KM Sagayam… - CCF Transactions on …, 2023 - Springer
The current milieu, encourages rapid growth of wireless communication, multimedia
applications, robotics and graphics to have efficient utilization of resources with high …

Towards a better 16-bit number representation for training neural networks

H De Silva, H Tan, NM Ho, JL Gustafson… - Conference on Next …, 2023 - Springer
Error resilience in neural networks has allowed for the adoption of low-precision floating-
point representations for mixed-precision training to improve efficiency. Although the IEEE …

Posit arithmetic for the training and deployment of generative adversarial networks

NM Ho, DT Nguyen, H De Silva… - … , Automation & Test …, 2021 - ieeexplore.ieee.org
This paper proposes a set of methods that enables low precision posit™ arithmetic to be
successfully used for the training of generative adversarial networks (GANs) with minimal …

An approach for matrix multiplication of 32-bit fixed point numbers by means of 16-bit SIMD instructions on DSP

I Safonov, A Kornilov, D Makienko - Electronics, 2022 - mdpi.com
Matrix multiplication is an important operation for many engineering applications.
Sometimes new features that include matrix multiplication should be added to existing and …

Physics-Informed Machine Learning for the Earth Sciences: Applications to Glaciology and Paleomagnetism

FF Sapienza - 2024 - search.proquest.com
This dissertation studies the application of machine learning in the fields of Glaciology and
Paleomagnetism. In the past few years, there have been significant advances in introducing …

[PDF][PDF] Exploring Segnet Architectures for iGPU Embedded Devices.

JB Chaudron, AM González - IJCCI, 2023 - scitepress.org
Image segmentation is an important topic in computer vision which encompasses a variety
of techniques to divide image into multiple areas or sub-regions in order to extract …

Randomization of approximate bilinear computation for matrix multiplication

OA Malik, S Becker - International Journal of Computer …, 2021 - Taylor & Francis
We present a method for randomizing formulas for bilinear computation of matrix products
which does not increase the leading order complexity of the computation. We consider the …

Towards a Better 16-Bit Number Representation for Training Neural Networks

WF Wong - … Conference, CoNGA 2023, Singapore, March 1-2 …, 2023 - books.google.com
Error resilience in neural networks has allowed for the adoption of low-precision floating-
point representations for mixed-precision training to improve efficiency. Although the IEEE …

Topics in Matrix and Tensor Computations

OA Malik - 2021 - search.proquest.com
This dissertation looks at matrices and tensors from a computational perspective. An
important problem in both matrix and tensor computations is decomposition. Several …