Knowledge Distillation and Gradient Estimation for Active Error Compensation in Approximate Neural Networks.
Research output: Contribution to book/Conference proceedings/Anthology/Report › Conference contribution › Contributed › peer-review
Contributors
Abstract
Approximate computing is a promising approach for optimizing computational resources of error-resilient applications such as Convolutional Neural Networks (CNNs). However, such approximations introduce an error that needs to be compensated by optimization methods, which typically include a retraining or fine-tuning stage. To efficiently recover from the introduced error, this fine-tuning process needs to be adapted to take CNN approximations into consideration. In this work, we present a novel methodology for fine-tuning approximate CNNs with ultralow bit-width quantization and large approximation error, which combines knowledge distillation and gradient estimation to recover the lost accuracy due to approximations. With our proposed methodology, we demonstrate energy savings of up to 38% in complex approximate CNNs with weights quantized to 4 bits and 8-bit activations, with less than 3% accuracy loss w.r.t. the full precision model.
Details
Original language | English |
---|---|
Title of host publication | Proceedings of the 2021 Design, Automation and Test in Europe, DATE 2021 |
Pages | 679-684 |
Number of pages | 6 |
ISBN (electronic) | 9783981926354 |
Publication status | Published - 1 Feb 2021 |
Peer-reviewed | Yes |
External IDs
Scopus | 85111030376 |
---|
Keywords
Research priority areas of TU Dresden
ASJC Scopus subject areas
Keywords
- Approximate Computing, Approximate multipliers, Neural Networks, Quantization