arXiv Open Access 2023

Assessing the Impact of Prompting Methods on ChatGPT's Mathematical Capabilities

Yuhao Chen Chloe Wong Hanwen Yang Juan Aguenza Sai Bhujangari +9 lainnya
Lihat Sumber

Abstrak

This study critically evaluates the efficacy of prompting methods in enhancing the mathematical reasoning capability of large language models (LLMs). The investigation uses three prescriptive prompting methods - simple, persona, and conversational prompting - known for their effectiveness in enhancing the linguistic tasks of LLMs. We conduct this analysis on OpenAI's LLM chatbot, ChatGPT-3.5, on extensive problem sets from the MATH, GSM8K, and MMLU datasets, encompassing a broad spectrum of mathematical challenges. A grading script adapted to each dataset is used to determine the effectiveness of these prompting interventions in enhancing the model's mathematical analysis power. Contrary to expectations, our empirical analysis reveals that none of the investigated methods consistently improves over ChatGPT-3.5's baseline performance, with some causing significant degradation. Our findings suggest that prompting strategies do not necessarily generalize to new domains, in this study failing to enhance mathematical performance.

Topik & Kata Kunci

Penulis (14)

Y

Yuhao Chen

C

Chloe Wong

H

Hanwen Yang

J

Juan Aguenza

S

Sai Bhujangari

B

Benthan Vu

X

Xun Lei

A

Amisha Prasad

M

Manny Fluss

E

Eric Phuong

M

Minghao Liu

R

Raja Kumar

V

Vanshika Vats

J

James Davis

Format Sitasi

Chen, Y., Wong, C., Yang, H., Aguenza, J., Bhujangari, S., Vu, B. et al. (2023). Assessing the Impact of Prompting Methods on ChatGPT's Mathematical Capabilities. https://arxiv.org/abs/2312.15006

Akses Cepat

Lihat di Sumber
Informasi Jurnal
Tahun Terbit
2023
Bahasa
en
Sumber Database
arXiv
Akses
Open Access ✓