Details
Presenter(s)
![Jian Huang Headshot](https://confcats-catavault.s3.amazonaws.com/CATAVault/ieeecass/master/files/styles/cc_user_photo/s3/user-pictures/13491.png?h=edbbed4d&itok=R8RP0Rf6)
Display Name
Jian Huang
- Affiliation
-
AffiliationNanjing University
- Country
Abstract
In this paper, an architecture is proposed for energy-efficient DNN training. It leverages triple sparsity in training which eliminates more unnecessary computation compared with prior works. Meanwhile, a 2-level grained mask-matching scheme is introduced for efficient sparsity detection. Coarse-grained mask match units are reused among PEs to save power and area while fine-grained mask match units are inside each PE to maintain throughput. As a result, our architecture achieves 42.1 TOPS and 174.0 TOPS/W in terms of throughput and energy efficiency, respectively. The energy efficiency of our design is 2.12× higher than the state-of-the-art training processor.