Knowledge Distillation and Gradient Estimation for Active Error Compensation in Approximate Neural Networks.

Research output: Contribution to book/Conference proceedings/Anthology/ReportConference contributionContributedpeer-review

Contributors

Abstract

Approximate computing is a promising approach for optimizing computational resources of error-resilient applications such as Convolutional Neural Networks (CNNs). However, such approximations introduce an error that needs to be compensated by optimization methods, which typically include a retraining or fine-tuning stage. To efficiently recover from the introduced error, this fine-tuning process needs to be adapted to take CNN approximations into consideration. In this work, we present a novel methodology for fine-tuning approximate CNNs with ultralow bit-width quantization and large approximation error, which combines knowledge distillation and gradient estimation to recover the lost accuracy due to approximations. With our proposed methodology, we demonstrate energy savings of up to 38% in complex approximate CNNs with weights quantized to 4 bits and 8-bit activations, with less than 3% accuracy loss w.r.t. the full precision model.

Details

Original languageEnglish
Title of host publicationProceedings of the 2021 Design, Automation and Test in Europe, DATE 2021
Pages679-684
Number of pages6
ISBN (electronic)9783981926354
Publication statusPublished - 1 Feb 2021
Peer-reviewedYes

External IDs

Scopus 85111030376

Keywords

Research priority areas of TU Dresden

ASJC Scopus subject areas

Keywords

  • Approximate Computing, Approximate multipliers, Neural Networks, Quantization