TY - GEN
T1 - XBM
T2 - 27th Asia and South Pacific Design Automation Conference, ASP-DAC 2022
AU - Zhang, Fan
AU - Yang, Li
AU - Meng, Jian
AU - Cao, Yu Kevin
AU - Seo, Jae Sun
AU - Fan, Deliang
N1 - Funding Information:
This work is supported in part by the National Science Foundation under Grant No.2003749 and No.1931871
Publisher Copyright:
© 2022 IEEE.
PY - 2022
Y1 - 2022
N2 - Recently, utilizing ReRAM crossbar array to accelerate DNN inference on single task has been widely studied. However, using the crossbar array for multiple task adaption has not been well explored. In this paper, for the first time, we propose XBM, a novel crossbar column-wise binary mask learning method for multiple task adaption in ReRAM crossbar DNN accelerator. XBM leverages the mask-based learning algorithm's benefit to avoid catastrophic forgetting to learn a task-specific mask for each new task. With our hardware-aware design innovation, the required masking operation to adapt for a new task could be easily implemented in existing crossbar based convolution engine with minimal hardware/ memory overhead and, more importantly, no need of power hungry cell re-programming, unlike prior works. The extensive experimental results show that compared with state-of-the-art multiple task adaption methods, XBM keeps the similar accuracy on new tasks while only requires 1.4% mask memory size compared with popular piggyback. Moreover, the elimination of cell re-programming or tuning saves up to 40% energy during new task adaption.
AB - Recently, utilizing ReRAM crossbar array to accelerate DNN inference on single task has been widely studied. However, using the crossbar array for multiple task adaption has not been well explored. In this paper, for the first time, we propose XBM, a novel crossbar column-wise binary mask learning method for multiple task adaption in ReRAM crossbar DNN accelerator. XBM leverages the mask-based learning algorithm's benefit to avoid catastrophic forgetting to learn a task-specific mask for each new task. With our hardware-aware design innovation, the required masking operation to adapt for a new task could be easily implemented in existing crossbar based convolution engine with minimal hardware/ memory overhead and, more importantly, no need of power hungry cell re-programming, unlike prior works. The extensive experimental results show that compared with state-of-the-art multiple task adaption methods, XBM keeps the similar accuracy on new tasks while only requires 1.4% mask memory size compared with popular piggyback. Moreover, the elimination of cell re-programming or tuning saves up to 40% energy during new task adaption.
UR - http://www.scopus.com/inward/record.url?scp=85126117584&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85126117584&partnerID=8YFLogxK
U2 - 10.1109/ASP-DAC52403.2022.9712508
DO - 10.1109/ASP-DAC52403.2022.9712508
M3 - Conference contribution
AN - SCOPUS:85126117584
T3 - Proceedings of the Asia and South Pacific Design Automation Conference, ASP-DAC
SP - 610
EP - 615
BT - ASP-DAC 2022 - 27th Asia and South Pacific Design Automation Conference, Proceedings
PB - Institute of Electrical and Electronics Engineers Inc.
Y2 - 17 January 2022 through 20 January 2022
ER -