Skip to main content
Video s3
    Details
    Poster
    Presenter(s)
    Hengrui Zhao Headshot
    Display Name
    Hengrui Zhao
    Affiliation
    Affiliation
    University of Science and Technology of China
    Country
    Country
    China
    Abstract

    In this paper, we propose to finetune and quantize a well-trained FPN convolutional network to obtain an integer convolutional network. Our key idea is to adjust the upper bound of a bounded rectified linear unit (ReLU), which replaces the normal ReLU and effectively controls the dynamic range of activations. Based on the tradeoff between learning ability and quantization error of networks, we managed to preserve full accuracy after quantization and obtain efficient integer networks. Our experiments on ResNet for image classification demonstrate that our 8-bit integer networks achieve state-of-the-art performance compared with Google's TensorFlow and NVIDIA's TensorRT. Moreover, we experiment on VDSR for image super-resolution and on VRCNN for compression artifact reduction, both of which serve regression tasks that natively require high inference accuracy. Besides ensuring the equivalent performance as the corresponding FPN networks, our integer networks have only 1/4 memory cost and run 2x faster on GPUs.

    Slides
    • Efficient Integer-Arithmetic-Only Convolutional Networks with Bounded ReLU (application/pdf)