Approximation opportunities in edge computing hardware: A systematic literature review
With the increasing popularity of the Internet of Things and massive Machine Type
Communication technologies, the number of connected devices is rising. However, although …
Communication technologies, the number of connected devices is rising. However, although …
Nimbus: Towards latency-energy efficient task offloading for ar services
Widespread adoption of mobile augmented reality (AR) and virtual reality (VR) applications
depends on their smoothness and immersiveness. Modern AR applications applying …
depends on their smoothness and immersiveness. Modern AR applications applying …
Graft: Efficient inference serving for hybrid deep learning with SLO guarantees via DNN re-alignment
Deep neural networks (DNNs) have been widely adopted for various mobile inference tasks,
yet their ever-increasing computational demands are hindering their deployment on …
yet their ever-increasing computational demands are hindering their deployment on …
Differentiable neural network pruning to enable smart applications on microcontrollers
Wearable, embedded, and IoT devices are a centrepiece of many ubiquitous computing
applications, such as fitness tracking, health monitoring, home security and voice assistants …
applications, such as fitness tracking, health monitoring, home security and voice assistants …
Special session: Towards an agile design methodology for efficient, reliable, and secure ML systems
The real-world use cases of Machine Learning (ML) have exploded over the past few years.
However, the current computing infrastructure is insufficient to support all real-world …
However, the current computing infrastructure is insufficient to support all real-world …
Hybrid slm and llm for edge-cloud collaborative inference
Edge-Cloud collaboration for deep learning inference has been actively studied, to enhance
the inference performance by leveraging both Edge and Cloud resources. However …
the inference performance by leveraging both Edge and Cloud resources. However …
Context-aware compilation of dnn training pipelines across edge and cloud
Empowered by machine learning, edge devices including smartphones, wearable, and IoT
devices have become growingly intelligent, raising conflicts with the limited resource. On …
devices have become growingly intelligent, raising conflicts with the limited resource. On …
Embedded Distributed Inference of Deep Neural Networks: A Systematic Review
FN Peccia, O Bringmann - arXiv preprint arXiv:2405.03360, 2024 - arxiv.org
Embedded distributed inference of Neural Networks has emerged as a promising approach
for deploying machine-learning models on resource-constrained devices in an efficient and …
for deploying machine-learning models on resource-constrained devices in an efficient and …
OODIn: An optimised on-device inference framework for heterogeneous mobile devices
SI Venieris, I Panopoulos… - 2021 IEEE International …, 2021 - ieeexplore.ieee.org
Radical progress in the field of deep learning (DL) has led to unprecedented accuracy in
diverse inference tasks. As such, deploying DL models across mobile platforms is vital to …
diverse inference tasks. As such, deploying DL models across mobile platforms is vital to …
A Survey on Securing Image-Centric Edge Intelligence
Facing enormous data generated at the network edge, Edge Intelligence (EI) emerges as
the fusion of Edge Computing and Artificial Intelligence, revolutionizing edge data …
the fusion of Edge Computing and Artificial Intelligence, revolutionizing edge data …