arXiv Open Access 2025

Hardware Efficient Accelerator for Spiking Transformer With Reconfigurable Parallel Time Step Computing

Bo-Yu Chen Tian-Sheuan Chang
Lihat Sumber

Abstrak

This paper introduces the first low-power hardware accelerator for Spiking Transformers, an emerging alternative to traditional artificial neural networks. By modifying the base Spikformer model to use IAND instead of residual addition, the model exclusively utilizes spike computation. The hardware employs a fully parallel tick-batching dataflow and a time-step reconfigurable neuron architecture, addressing the delay and power challenges of multi-timestep processing in spiking neural networks. This approach processes outputs from all time steps in parallel, reducing computation delay and eliminating membrane memory, thereby lowering energy consumption. The accelerator supports 3x3 and 1x1 convolutions and matrix operations through vectorized processing, meeting model requirements. Implemented in TSMC's 28nm process, it achieves 3.456 TSOPS (tera spike operations per second) with a power efficiency of 38.334 TSOPS/W at 500MHz, using 198.46K logic gates and 139.25KB of SRAM.

Topik & Kata Kunci

Penulis (2)

B

Bo-Yu Chen

T

Tian-Sheuan Chang

Format Sitasi

Chen, B., Chang, T. (2025). Hardware Efficient Accelerator for Spiking Transformer With Reconfigurable Parallel Time Step Computing. https://arxiv.org/abs/2503.19643

Akses Cepat

Lihat di Sumber
Informasi Jurnal
Tahun Terbit
2025
Bahasa
en
Sumber Database
arXiv
Akses
Open Access ✓