PEPPER: Profiling-based Edge Placement and Partitioning for Deep Learning Execution
Authors/Creators
Description
Unlocking the full potential of AI at the edge requires overcoming the fundamental challenge of running complex models efficiently on devices with limited computational power. In this work, the challenge of optimizing the deployment of deep learning models in resource-constrained environments is addressed. A novel pipeline is proposed for profiling and partitioning ONNX models, to enhance inference efficiency across heterogeneous hardware platforms. Optimal split points within the deep learning models are identified through the application of Tarjan’s Bridge-Finding Algorithm, and the inference times of the models are predicted per device based on the respective characteristics and CPU load. For the prediction of inference times, the XGBoost algorithm is employed. The effectiveness of the proposed approach is validated through experiments conducted on real-world edge devices, demonstrating that highly efficient and adaptable deployment of complex deep learning models can be achieved in such environments.
Files
PROfiling_based_Edge_Placement_and_Partitioning_for_DL_Execution (pre-print).pdf
Files
(796.4 kB)
| Name | Size | Download all |
|---|---|---|
|
md5:df1a1183e28961e777e1d7e974e31192
|
796.4 kB | Preview Download |