arXiv Open Access 2024

DeMansia: Mamba Never Forgets Any Tokens

Ricky Fang
Lihat Sumber

Abstrak

This paper examines the mathematical foundations of transformer architectures, highlighting their limitations particularly in handling long sequences. We explore prerequisite models such as Mamba, Vision Mamba (ViM), and LV-ViT that pave the way for our proposed architecture, DeMansia. DeMansia integrates state space models with token labeling techniques to enhance performance in image classification tasks, efficiently addressing the computational challenges posed by traditional transformers. The architecture, benchmark, and comparisons with contemporary models demonstrate DeMansia's effectiveness. The implementation of this paper is available on GitHub at https://github.com/catalpaaa/DeMansia

Topik & Kata Kunci

Penulis (1)

R

Ricky Fang

Format Sitasi

Fang, R. (2024). DeMansia: Mamba Never Forgets Any Tokens. https://arxiv.org/abs/2408.01986

Akses Cepat

Lihat di Sumber
Informasi Jurnal
Tahun Terbit
2024
Bahasa
en
Sumber Database
arXiv
Akses
Open Access ✓