Semantic Scholar Open Access 2023 701 sitasi

A Comprehensive Survey on Pretrained Foundation Models: A History from BERT to ChatGPT

Ce Zhou Qian Li Chen Li Jun Yu Yixin Liu +22 lainnya

Abstrak

Pretrained Foundation Models (PFMs) are regarded as the foundation for various downstream tasks with different data modalities. A PFM (e.g., BERT, ChatGPT, and GPT-4) is trained on large-scale data which provides a reasonable parameter initialization for a wide range of downstream applications. BERT learns bidirectional encoder representations from Transformers, which are trained on large datasets as contextual language models. Similarly, the generative pretrained transformer (GPT) method employs Transformers as the feature extractor and is trained using an autoregressive paradigm on large datasets. Recently, ChatGPT shows promising success on large language models, which applies an autoregressive language model with zero shot or few shot prompting. The remarkable achievements of PFM have brought significant breakthroughs to various fields of AI. Numerous studies have proposed different methods, raising the demand for an updated survey. This study provides a comprehensive review of recent research advancements, challenges, and opportunities for PFMs in text, image, graph, as well as other data modalities. The review covers the basic components and existing pretraining methods used in natural language processing, computer vision, and graph learning. Additionally, it explores advanced PFMs used for different data modalities and unified PFMs that consider data quality and quantity. The review also discusses research related to the fundamentals of PFMs, such as model efficiency and compression, security, and privacy. Finally, the study provides key implications, future research directions, challenges, and open problems in the field of PFMs. Overall, this survey aims to shed light on the research of the PFMs on scalability, security, logical reasoning ability, cross-domain learning ability, and the user-friendly interactive ability for artificial general intelligence.

Topik & Kata Kunci

Penulis (27)

C

Ce Zhou

Q

Qian Li

C

Chen Li

J

Jun Yu

Y

Yixin Liu

G

Guangjing Wang

K

Kaichao Zhang

C

Cheng Ji

Q

Qi Yan

L

Lifang He

H

Hao Peng

J

Jianxin Li

J

Jia Wu

Z

Ziwei Liu

P

Pengtao Xie

C

Caiming Xiong

J

Jian Pei

P

Philip S. Yu

L

Lichao Sun Michigan State University

B

B. University

L

Lehigh University

M

Macquarie University

N

Nanyang Technological University

U

University of California at San Diego

D

Duke University

U

U. Chicago

S

S. Research

Format Sitasi

Zhou, C., Li, Q., Li, C., Yu, J., Liu, Y., Wang, G. et al. (2023). A Comprehensive Survey on Pretrained Foundation Models: A History from BERT to ChatGPT. https://doi.org/10.48550/arXiv.2302.09419

Akses Cepat

Lihat di Sumber doi.org/10.48550/arXiv.2302.09419
Informasi Jurnal
Tahun Terbit
2023
Bahasa
en
Total Sitasi
701×
Sumber Database
Semantic Scholar
DOI
10.48550/arXiv.2302.09419
Akses
Open Access ✓