arXiv Open Access 2025

Mamba-X: An End-to-End Vision Mamba Accelerator for Edge Computing Devices

Dongho Yoon Gungyu Lee Jaewon Chang Yunjae Lee Dongjae Lee +1 lainnya
Lihat Sumber

Abstrak

Transformers have proven effective in language modeling but are limited by high computational and memory demands that grow quadratically with input sequence length. State space models (SSMs) offer a promising alternative by reducing attention complexity from $O(L^2)$ to $O(L)$ while also lowering overall memory consumption. Vision Mamba adapts the SSM approach for computer vision tasks, achieving lower latency and memory consumption than traditional transformer models. However, deploying Vision Mamba on edge devices is challenging due to its sequential scan operations, which hinder GPU efficiency. We propose Mamba-X, an end-to-end Vision Mamba accelerator that includes a systolic scan array to maximize parallelism and minimize memory traffic, along with a hybrid, hardware-friendly quantization technique to reduce memory usage and improve hardware efficiency without sacrificing accuracy.

Topik & Kata Kunci

Penulis (6)

D

Dongho Yoon

G

Gungyu Lee

J

Jaewon Chang

Y

Yunjae Lee

D

Dongjae Lee

M

Minsoo Rhu

Format Sitasi

Yoon, D., Lee, G., Chang, J., Lee, Y., Lee, D., Rhu, M. (2025). Mamba-X: An End-to-End Vision Mamba Accelerator for Edge Computing Devices. https://arxiv.org/abs/2508.02977

Akses Cepat

Lihat di Sumber
Informasi Jurnal
Tahun Terbit
2025
Bahasa
en
Sumber Database
arXiv
Akses
Open Access ✓