LLaMA: Open and Efficient Foundation Language Models
Abstrak
We introduce LLaMA, a collection of foundation language models ranging from 7B to 65B parameters. We train our models on trillions of tokens, and show that it is possible to train state-of-the-art models using publicly available datasets exclusively, without resorting to proprietary and inaccessible datasets. In particular, LLaMA-13B outperforms GPT-3 (175B) on most benchmarks, and LLaMA-65B is competitive with the best models, Chinchilla-70B and PaLM-540B. We release all our models to the research community.
Topik & Kata Kunci
Penulis (14)
Hugo Touvron
Thibaut Lavril
Gautier Izacard
X. Martinet
M. Lachaux
Timothée Lacroix
Baptiste Rozière
Naman Goyal
Eric Hambro
Faisal Azhar
Aur'elien Rodriguez
Armand Joulin
Edouard Grave
Guillaume Lample
Akses Cepat
PDF tidak tersedia langsung
Cek di sumber asli →- Tahun Terbit
- 2023
- Bahasa
- en
- Total Sitasi
- 19293×
- Sumber Database
- Semantic Scholar
- Akses
- Open Access ✓