Performance models and energy-optimal scheduling of DNNs on many-core hardware with dynamic power management

Research output: Contribution to book/Conference proceedings/Anthology/ReportConference contributionContributedpeer-review

Abstract

Processing of deep neural networks (DNNs) at the edge may be limited by power or energy constraints of the used embedded hardware system. It is therefore desirable for the compiler to create efficient executables for given DNN models meeting the specific constraints. Here, we consider a low-power many-core hardware with 152 processing elements (PE), each containing an ARM M4F processor, 128 KB SRAM and a custom accelerator for DNN inference. Dynamic power management allows each core to switch between a high-speed and a low-power mode within tens of nanoseconds. For an energy-optimal parallelization of DNNs on the hardware, we first develop analytical performance models to predict the time and energy for executing a DNN layer with the custom accelerator. The models are fitted and validated using measurements on a prototype chip. In a second step we develop concepts for the energy-optimal parallelization of DNNs under latency constraints and evaluate them deploying the performance models: By dynamically switching between the operating modes more than 10% of energy can be saved compared to the case running at high-speed mode only. The presented methodology and concepts are easily transferable to other many-core edge processors.

Details

Original languageEnglish
Title of host publicationProceedings - 2023 IEEE/ACM International Workshop on Compilers, Deployment, and Tooling for Edge AI, CODAI 2023
Pages27 - 31
Number of pages5
ISBN (electronic)9798400703379
Publication statusPublished - 21 Sept 2023
Peer-reviewedYes

External IDs

ORCID /0000-0002-6286-5064/work/166324419
Scopus 85196429407

Keywords

Keywords

  • deep neural networks, edge computing, many-core hardware, parallelization, performance model, power management