arXiv Open Access 2022

MIME: Adapting a Single Neural Network for Multi-task Inference with Memory-efficient Dynamic Pruning

Abhiroop Bhattacharjee Yeshwanth Venkatesha Abhishek Moitra Priyadarshini Panda
Lihat Sumber

Abstrak

Recent years have seen a paradigm shift towards multi-task learning. This calls for memory and energy-efficient solutions for inference in a multi-task scenario. We propose an algorithm-hardware co-design approach called MIME. MIME reuses the weight parameters of a trained parent task and learns task-specific threshold parameters for inference on multiple child tasks. We find that MIME results in highly memory-efficient DRAM storage of neural-network parameters for multiple tasks compared to conventional multi-task inference. In addition, MIME results in input-dependent dynamic neuronal pruning, thereby enabling energy-efficient inference with higher throughput on a systolic-array hardware. Our experiments with benchmark datasets (child tasks)- CIFAR10, CIFAR100, and Fashion-MNIST, show that MIME achieves ~3.48x memory-efficiency and ~2.4-3.1x energy-savings compared to conventional multi-task inference in Pipelined task mode.

Topik & Kata Kunci

Penulis (4)

A

Abhiroop Bhattacharjee

Y

Yeshwanth Venkatesha

A

Abhishek Moitra

P

Priyadarshini Panda

Format Sitasi

Bhattacharjee, A., Venkatesha, Y., Moitra, A., Panda, P. (2022). MIME: Adapting a Single Neural Network for Multi-task Inference with Memory-efficient Dynamic Pruning. https://arxiv.org/abs/2204.05274

Akses Cepat

Lihat di Sumber
Informasi Jurnal
Tahun Terbit
2022
Bahasa
en
Sumber Database
arXiv
Akses
Open Access ✓