arXiv Open Access 2023

MUFFIN: Curating Multi-Faceted Instructions for Improving Instruction-Following

Renze Lou Kai Zhang Jian Xie Yuxuan Sun Janice Ahn +3 lainnya
Lihat Sumber

Abstrak

In the realm of large language models (LLMs), enhancing instruction-following capability often involves curating expansive training data. This is achieved through two primary schemes: i) Scaling-Inputs: Amplifying (input, output) pairs per task instruction, aiming for better instruction adherence. ii) Scaling Input-Free Tasks: Enlarging tasks, each composed of an (instruction, output) pair (without requiring a separate input anymore). However, LLMs under Scaling-Inputs tend to be overly sensitive to inputs, leading to misinterpretation or non-compliance with instructions. Conversely, Scaling Input-Free Tasks demands a substantial number of tasks but is less effective in instruction following when dealing with instances in Scaling-Inputs. This work introduces MUFFIN, a new scheme of instruction-following dataset curation. Specifically, we automatically Scale Tasks per Input by diversifying these tasks with various input facets. Experimental results across four zero-shot benchmarks, spanning both Scaling-Inputs and Scaling Input-Free Tasks schemes, reveal that LLMs, at various scales, trained on MUFFIN generally demonstrate superior instruction-following capabilities compared to those trained on the two aforementioned schemes.

Topik & Kata Kunci

Penulis (8)

R

Renze Lou

K

Kai Zhang

J

Jian Xie

Y

Yuxuan Sun

J

Janice Ahn

H

Hanzi Xu

Y

Yu Su

W

Wenpeng Yin

Format Sitasi

Lou, R., Zhang, K., Xie, J., Sun, Y., Ahn, J., Xu, H. et al. (2023). MUFFIN: Curating Multi-Faceted Instructions for Improving Instruction-Following. https://arxiv.org/abs/2312.02436

Akses Cepat

Lihat di Sumber
Informasi Jurnal
Tahun Terbit
2023
Bahasa
en
Sumber Database
arXiv
Akses
Open Access ✓