arXiv Open Access 2024

Kajal: Extracting Grammar of a Source Code Using Large Language Models

Mohammad Jalili Torkamani
Lihat Sumber

Abstrak

Understanding and extracting the grammar of a domain-specific language (DSL) is crucial for various software engineering tasks; however, manually creating these grammars is time-intensive and error-prone. This paper presents Kajal, a novel approach that automatically infers grammar from DSL code snippets by leveraging Large Language Models (LLMs) through prompt engineering and few-shot learning. Kajal dynamically constructs input prompts, using contextual information to guide the LLM in generating the corresponding grammars, which are iteratively refined through a feedback-driven approach. Our experiments show that Kajal achieves 60% accuracy with few-shot learning and 45% without it, demonstrating the significant impact of few-shot learning on the tool's effectiveness. This approach offers a promising solution for automating DSL grammar extraction, and future work will explore using smaller, open-source LLMs and testing on larger datasets to further validate Kajal's performance.

Topik & Kata Kunci

Penulis (1)

M

Mohammad Jalili Torkamani

Format Sitasi

Torkamani, M.J. (2024). Kajal: Extracting Grammar of a Source Code Using Large Language Models. https://arxiv.org/abs/2412.08842

Akses Cepat

Lihat di Sumber
Informasi Jurnal
Tahun Terbit
2024
Bahasa
en
Sumber Database
arXiv
Akses
Open Access ✓