A review of emerging trends in photonic deep learning accelerators

M Atwany, S Pardo, S Serunjogi, M Rasras - Frontiers in Physics, 2024 - frontiersin.org
Deep learning has revolutionized many sectors of industry and daily life, but as application
scale increases, performing training and inference with large models on massive datasets is …

Fsa: An efficient fault-tolerant systolic array-based dnn accelerator architecture

Y Zhao, K Wang, A Louri - 2022 IEEE 40th International …, 2022 - ieeexplore.ieee.org
With the advent of Deep Neural Network (DNN) accelerators, permanent faults are
increasingly becoming a serious challenge for DNN hardware accelerator, as they can …

SwInt: A Non-Blocking Switch-Based Silicon Photonic Interposer Network for 2.5 D Machine Learning Accelerators

E Taheri, MA Mahdian, S Pasricha… - IEEE Journal on …, 2024 - ieeexplore.ieee.org
The surging demand for machine learning (ML) applications has emphasized the pressing
need for efficient ML accelerators capable of addressing the computational and energy …

Big-little chiplets for in-memory acceleration of dnns: A scalable heterogeneous architecture

G Krishnan, AA Goksoy, SK Mandal, Z Wang… - Proceedings of the 41st …, 2022 - dl.acm.org
Monolithic in-memory computing (IMC) architectures face significant yield and fabrication
cost challenges as the complexity of DNNs increases. Chiplet-based IMCs that integrate …

A High-Performance and Energy-Efficient Photonic Architecture for Multi-DNN Acceleration

Y Li, A Louri, A Karanth - IEEE Transactions on Parallel and …, 2023 - ieeexplore.ieee.org
Large-scale deep neural network (DNN) accelerators are poised to facilitate the concurrent
processing of diverse DNNs, imposing demanding challenges on the interconnection fabric …

Locmoe: A low-overhead moe for large language model training

J Li, Z Sun, X He, L Zeng, Y Lin, E Li, B Zheng… - arXiv preprint arXiv …, 2024 - arxiv.org
The Mixtures-of-Experts (MoE) model is a widespread distributed and integrated learning
method for large language models (LLM), which is favored due to its ability to sparsify and …

Trade-off-oriented impedance optimization of chiplet-based 2.5-D integrated circuits with a hybrid MDP algorithm for noise elimination

C Zhi, G Dong, Y Wang, Z Zhu… - IEEE Transactions on …, 2022 - ieeexplore.ieee.org
Interposer and chiplet-based 2.5-D integrated circuit (IC) designs have become a new trend
for block-level heterogeneous integration. In this paper, a new hybrid metaheuristic …

A silicon photonic multi-DNN accelerator

Y Li, A Louri, A Karanth - 2023 32nd International Conference …, 2023 - ieeexplore.ieee.org
In shared environments like cloud-based datacenters, hardware accelerators are deployed
to meet the scale-out computation demands of deep neural network (DNN) inference tasks …

SEECHIP: A Scalable and Energy-Efficient Chiplet-based GPU Architecture Using Photonic Links

H Zhang, Y Chen, Z Huang, H Zhang… - Proceedings of the 52nd …, 2023 - dl.acm.org
The continuous increase in GPU performance benefits a wide range of high-performance
computing (HPC) applications. Slower growth of transistor density and limited size of chip …

[HTML][HTML] ChipAI: A scalable chiplet-based accelerator for efficient DNN inference using silicon photonics

H Zhang, H Zhang, Z Huang, Y Chen - Journal of Systems Architecture, 2025 - Elsevier
To enhance the precision of inference, deep neural network (DNN) models have been
progressively growing in scale and complexity, leading to increased latency and …