Skip to main content
Video s3
    Details
    Poster
    Presenter(s)
    Jian Meng Headshot
    Display Name
    Jian Meng
    Affiliation
    Affiliation
    Arizona State University
    Country
    Abstract

    The high computational complexity and a large number of parameters of deep neural networks become the most intensive burden of deep learning hardware design. The resistive RAM (RRAM) crossbar based in-memory computing has recently emerged as a promising DNN acceleration technique. To fully exploit crossbar-based IMC efficiency, a systematic software and hardware design for compression are necessary. This paper presents a system-level design considering the low precision quantization, structured pruning, and RRAM crossbar mapping. Both hardware and software results of the proposed design show the large performance improvements and superiority compared with the recent state-of-art compression design.

    Slides
    • Structured Pruning of RRAM Crossbars for Efficient In-Memory Computing Acceleration of Deep Neural Networks (application/pdf)