Skip to main content
Video s3
    Details
    Presenter(s)
    Danilo Vucetic Headshot
    Display Name
    Danilo Vucetic
    Affiliation
    Affiliation
    McGill University
    Country
    Country
    Canada
    Author(s)
    Display Name
    Danilo Vucetic
    Affiliation
    Affiliation
    McGill University
    Affiliation
    Affiliation
    McGill University
    Display Name
    Maryam Ziaeefard
    Affiliation
    Affiliation
    McGill University
    Display Name
    James J. Clark
    Affiliation
    Affiliation
    McGill University
    Display Name
    Brett H. Meyer
    Affiliation
    Affiliation
    McGill University
    Display Name
    Warren Gross
    Affiliation
    Affiliation
    McGill University
    Abstract

    We propose a method to reduce the fine-tuning costs of BERT-based natural language models such that they may be realized on resource-constrained devices. We identify memory usage as a major bottleneck and reduce memory operations during fine-tuning by training only a subset of the model. A reconfiguration of the model achieves better memory performance and training time. Our approach reduces memory usage, memory access time, and fine-tuning time substantially, while achieving near-baseline metric performance.

    Slides
    • Efficient Fine-Tuning of BERT Models on the Edge (application/pdf)