arXiv Open Access 2025

SafeRL-Lite: A Lightweight, Explainable, and Constrained Reinforcement Learning Library

Satyam Mishra Phung Thao Vi Shivam Mishra Vishwanath Bijalwan Vijay Bhaskar Semwal +1 lainnya
Lihat Sumber

Abstrak

We introduce SafeRL-Lite, an open-source Python library for building reinforcement learning (RL) agents that are both constrained and explainable. Existing RL toolkits often lack native mechanisms for enforcing hard safety constraints or producing human-interpretable rationales for decisions. SafeRL-Lite provides modular wrappers around standard Gym environments and deep Q-learning agents to enable: (i) safety-aware training via constraint enforcement, and (ii) real-time post-hoc explanation via SHAP values and saliency maps. The library is lightweight, extensible, and installable via pip, and includes built-in metrics for constraint violations. We demonstrate its effectiveness on constrained variants of CartPole and provide visualizations that reveal both policy logic and safety adherence. The full codebase is available at: https://github.com/satyamcser/saferl-lite.

Topik & Kata Kunci

Penulis (6)

S

Satyam Mishra

P

Phung Thao Vi

S

Shivam Mishra

V

Vishwanath Bijalwan

V

Vijay Bhaskar Semwal

A

Abdul Manan Khan

Format Sitasi

Mishra, S., Vi, P.T., Mishra, S., Bijalwan, V., Semwal, V.B., Khan, A.M. (2025). SafeRL-Lite: A Lightweight, Explainable, and Constrained Reinforcement Learning Library. https://arxiv.org/abs/2506.17297

Akses Cepat

Lihat di Sumber
Informasi Jurnal
Tahun Terbit
2025
Bahasa
en
Sumber Database
arXiv
Akses
Open Access ✓