TY - JOUR
T1 - Everything Leaves Footprints
T2 - Hardware Accelerated Intermittent Deep Inference
AU - Kang, Chih Kai
AU - Mendis, Hashan Roshantha
AU - Lin, Chun Han
AU - Chen, Ming Syan
AU - Hsiu, Pi Cheng
N1 - Funding Information:
Manuscript received March 26, 2020; revised June 9, 2020; accepted July 6, 2020. Date of publication October 2, 2020; date of current version October 27, 2020. This work was supported in part by the Project for Excellent Junior Research Investigators, Ministry of Science and Technology, Taiwan, under Grant MOST 107-2628-E-001-001-MY3. This article was presented in the International Conference on Hardware/Software Codesign and System Synthesis 2020 and appears as part of the ESWEEK-TCAD special issue. (Corresponding author: Pi-Cheng Hsiu.) Chih-Kai Kang and Ming-Syan Chen are with the Graduate Institute of Electrical Engineering, National Taiwan University, Taipei 10617, Taiwan, and also with the Research Center for Information Technology Innovation, Academia Sinica, Taipei 11529, Taiwan (e-mail: ckkang@arbor.ee.ntu.edu.tw; mschen@ntu.edu.tw).
Publisher Copyright:
© 1982-2012 IEEE.
PY - 2020/11
Y1 - 2020/11
N2 - Current peripheral execution approaches for intermittently powered systems require full access to the internal hardware state for checkpointing or rely on application-level energy estimation for task partitioning to make correct forward progress. Both requirements present significant practical challenges for energy-harvesting, intelligent edge Internet-of-Things devices, which perform hardware-accelerated deep neural network (DNN) inference. Sophisticated compute peripherals may have an inaccessible internal state, and the complexity of DNN models makes it difficult for programmers to partition the application into suitably sized tasks that fit within an estimated energy budget. This article presents the concept of inference footprinting for intermittent DNN inference, where accelerator progress is accumulatively preserved across power cycles. Our middleware stack, HAWAII, tracks and restores inference footprints efficiently and transparently to make inference forward progress, without requiring access to the accelerator internal state and application-level energy estimation. Evaluations were carried out on a Texas Instruments device, under varied energy budgets and network workloads. Compared to a variety of task-based intermittent approaches, HAWAII improves the inference throughput by 5.7%-95.7%, particularly achieving higher performance on heavily accelerated DNNs.
AB - Current peripheral execution approaches for intermittently powered systems require full access to the internal hardware state for checkpointing or rely on application-level energy estimation for task partitioning to make correct forward progress. Both requirements present significant practical challenges for energy-harvesting, intelligent edge Internet-of-Things devices, which perform hardware-accelerated deep neural network (DNN) inference. Sophisticated compute peripherals may have an inaccessible internal state, and the complexity of DNN models makes it difficult for programmers to partition the application into suitably sized tasks that fit within an estimated energy budget. This article presents the concept of inference footprinting for intermittent DNN inference, where accelerator progress is accumulatively preserved across power cycles. Our middleware stack, HAWAII, tracks and restores inference footprints efficiently and transparently to make inference forward progress, without requiring access to the accelerator internal state and application-level energy estimation. Evaluations were carried out on a Texas Instruments device, under varied energy budgets and network workloads. Compared to a variety of task-based intermittent approaches, HAWAII improves the inference throughput by 5.7%-95.7%, particularly achieving higher performance on heavily accelerated DNNs.
KW - Deep neural networks (DNNs)
KW - edge computing
KW - energy harvesting
KW - intermittent systems
UR - http://www.scopus.com/inward/record.url?scp=85096034375&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85096034375&partnerID=8YFLogxK
U2 - 10.1109/TCAD.2020.3012217
DO - 10.1109/TCAD.2020.3012217
M3 - Article
AN - SCOPUS:85096034375
SN - 0278-0070
VL - 39
SP - 3479
EP - 3491
JO - IEEE Transactions on Computer-Aided Design of Integrated Circuits and Systems
JF - IEEE Transactions on Computer-Aided Design of Integrated Circuits and Systems
IS - 11
M1 - 9211553
ER -