CrossRef Open Access 2025

AlphaViT: a flexible game-playing AI for multiple games and variable board sizes

Kazuhisa Fujita

Abstrak

This study presents three game-playing agents incorporating Vision Transformers (ViTs) into the AlphaZero framework: AlphaViT, AlphaViD (AlphaViT with a transformer decoder), and AlphaVDA (AlphaViD with learnable action embeddings). These agents can play multiple board games with varying sizes using a single shared-weight neural network, thus overcoming AlphaZero’s limitation of fixed board sizes. AlphaViT employs only a transformer encoder, whereas AlphaViD and AlphaVDA incorporate both a transformer encoder and a decoder. In AlphaViD, the decoder processes the output from the encoder, whereas AlphaVDA uses learnable embeddings as decoder input. The additional decoder in AlphaViD and AlphaVDA provides flexibility to adapt to various action spaces and board sizes. Experimental results show that the proposed agents, trained on either individual games or multiple games simultaneously, consistently outperform traditional algorithms, such as Minimax and Monte Carlo Tree Search. They achieve performance close to that of AlphaZero despite relying on a single deep neural network (DNN) with shared weights. In particular, AlphaViT performs well across all evaluated games. Furthermore, fine-tuning the DNN using weights pre-trained on small board games accelerates convergence and improves performance, particularly in Gomoku. Notably, simultaneous training on multiple games yields performance comparable to, or even surpassing, that of single-game training. These results indicate the potential of transformer-based architectures for developing flexible and robust game-playing AI agents that excel in multiple games and dynamic environments.

Penulis (1)

K

Kazuhisa Fujita

Format Sitasi

Fujita, K. (2025). AlphaViT: a flexible game-playing AI for multiple games and variable board sizes. https://doi.org/10.7717/peerj-cs.3403

Akses Cepat

Lihat di Sumber doi.org/10.7717/peerj-cs.3403
Informasi Jurnal
Tahun Terbit
2025
Bahasa
en
Sumber Database
CrossRef
DOI
10.7717/peerj-cs.3403
Akses
Open Access ✓