Skip to main content
Video s3
    Details
    Author(s)
    Display Name
    Tianyi Zhang
    Affiliation
    Affiliation
    Arizona State University
    Affiliation
    Affiliation
    Arizona State University
    Display Name
    Dong-Woo Jee
    Affiliation
    Affiliation
    Ajou University
    Display Name
    Injune Yeo
    Affiliation
    Affiliation
    Arizona State University
    Display Name
    Yaoxin Zhuo
    Affiliation
    Affiliation
    Arizona State University
    Display Name
    Baoxin Li
    Affiliation
    Affiliation
    Arizona State University
    Display Name
    Mark Rodder
    Affiliation
    Affiliation
    Samsung Semiconductor Inc.
    Display Name
    Yu Cao
    Affiliation
    Affiliation
    Arizona State University
    Abstract

    The resolution of CMOS image sensors keeps increasing, posing a fundamental challenge to sensor throughput and efficiency. Inspired by selective attention in human vision, we introduce a saliency step to continuously select salient pixels, and reduce output volume, power consumption and latency. To minimize the overhead, we integrate three methods: image down-sampling, model reduction, and finally padding at a minimum for the same accuracy of object detection with selected pixels. We demonstrate our approach on various datasets, achieving 70.5% reduction in the output volume for BDD100K, which translates to 4.3× and 3.4× reduction in power consumption and latency, respectively.