A Survey on Design Methodologies for Accelerating Deep Learning on Heterogeneous Architectures
Abstrak
Given their increasing size and complexity, the need for efficient execution of deep neural networks has become increasingly pressing in the design of heterogeneous High-Performance Computing (HPC) and edge platforms, leading to a wide variety of proposals for specialized deep learning architectures and hardware accelerators. The design of such architectures and accelerators requires a multidisciplinary approach combining expertise from several areas, from machine learning to computer architecture, low-level hardware design, and approximate computing. Several methodologies and tools have been proposed to improve the process of designing accelerators for deep learning, aimed at maximizing parallelism and minimizing data movement to achieve high performance and energy efficiency. This paper critically reviews influential tools and design methodologies for Deep Learning accelerators, offering a wide perspective in this rapidly evolving field. This work complements surveys on architectures and accelerators by covering hardware-software co-design, automated synthesis, domain-specific compilers, design space exploration, modeling, and simulation, providing insights into technical challenges and open research directions.
Penulis (18)
Serena Curzel
Fabrizio Ferrandi
Leandro Fiorin
Daniele Ielmini
Cristina Silvano
Francesco Conti
Luca Bompani
Luca Benini
Enrico Calore
Sebastiano Fabio Schifano
Cristian Zambelli
Maurizio Palesi
Giuseppe Ascia
Enrico Russo
Valeria Cardellini
Salvatore Filippone
Francesco Lo Presti
Stefania Perri
Akses Cepat
- Tahun Terbit
- 2023
- Bahasa
- en
- Sumber Database
- arXiv
- Akses
- Open Access ✓