Poster: Energy-Aware Partitioning for Edge AI

More Info
expand_more

Abstract

Model partitioning is a promising solution to reduce the high computation load and transmission of high-volume data. Within the scope of Edge AI, the fundamentals of model partitioning involve splitting the model for local computing at the edge and offloading heavy computation tasks to the cloud or server. This approach benefits scenarios with limited computing and battery capacity with low latency requirements, such as connected autonomous vehicles. However, while model partitioning offers advantages in reducing the onboard computation, memory requirements and inference time, it also introduces challenges such as increased energy consumption for partitioned computations and overhead for transferring partitioned data/model. In this work, we explore hybrid model partitioning to optimize computational and communication energy consumption. Our results provide an initial analysis of the tradeoff between energy and accuracy, focusing on the energy-aware model partitioning for future Edge AI applications.

Files

Poster_Energy-Aware_Partitioni... (pdf)
(pdf | 0.363 Mb)
Unknown license
warning

File under embargo until 01-07-2025