arXiv Open Access 2024

A Benchmark for Long-Form Medical Question Answering

Pedram Hosseini Jessica M. Sin Bing Ren Bryceton G. Thomas Elnaz Nouri +2 lainnya
Lihat Sumber

Abstrak

There is a lack of benchmarks for evaluating large language models (LLMs) in long-form medical question answering (QA). Most existing medical QA evaluation benchmarks focus on automatic metrics and multiple-choice questions. While valuable, these benchmarks fail to fully capture or assess the complexities of real-world clinical applications where LLMs are being deployed. Furthermore, existing studies on evaluating long-form answer generation in medical QA are primarily closed-source, lacking access to human medical expert annotations, which makes it difficult to reproduce results and enhance existing baselines. In this work, we introduce a new publicly available benchmark featuring real-world consumer medical questions with long-form answer evaluations annotated by medical doctors. We performed pairwise comparisons of responses from various open and closed-source medical and general-purpose LLMs based on criteria such as correctness, helpfulness, harmfulness, and bias. Additionally, we performed a comprehensive LLM-as-a-judge analysis to study the alignment between human judgments and LLMs. Our preliminary results highlight the strong potential of open LLMs in medical QA compared to leading closed models. Code & Data: https://github.com/lavita-ai/medical-eval-sphere

Topik & Kata Kunci

Penulis (7)

P

Pedram Hosseini

J

Jessica M. Sin

B

Bing Ren

B

Bryceton G. Thomas

E

Elnaz Nouri

A

Ali Farahanchi

S

Saeed Hassanpour

Format Sitasi

Hosseini, P., Sin, J.M., Ren, B., Thomas, B.G., Nouri, E., Farahanchi, A. et al. (2024). A Benchmark for Long-Form Medical Question Answering. https://arxiv.org/abs/2411.09834

Akses Cepat

Lihat di Sumber
Informasi Jurnal
Tahun Terbit
2024
Bahasa
en
Sumber Database
arXiv
Akses
Open Access ✓