arXiv Open Access 2021

Differentially Private Fine-tuning of Language Models

Da Yu Saurabh Naik Arturs Backurs Sivakanth Gopi Huseyin A. Inan +7 lainnya
Lihat Sumber

Abstrak

We give simpler, sparser, and faster algorithms for differentially private fine-tuning of large-scale pre-trained language models, which achieve the state-of-the-art privacy versus utility tradeoffs on many standard NLP tasks. We propose a meta-framework for this problem, inspired by the recent success of highly parameter-efficient methods for fine-tuning. Our experiments show that differentially private adaptations of these approaches outperform previous private algorithms in three important dimensions: utility, privacy, and the computational and memory cost of private training. On many commonly studied datasets, the utility of private models approaches that of non-private models. For example, on the MNLI dataset we achieve an accuracy of $87.8\%$ using RoBERTa-Large and $83.5\%$ using RoBERTa-Base with a privacy budget of $ε= 6.7$. In comparison, absent privacy constraints, RoBERTa-Large achieves an accuracy of $90.2\%$. Our findings are similar for natural language generation tasks. Privately fine-tuning with DART, GPT-2-Small, GPT-2-Medium, GPT-2-Large, and GPT-2-XL achieve BLEU scores of 38.5, 42.0, 43.1, and 43.8 respectively (privacy budget of $ε= 6.8,δ=$ 1e-5) whereas the non-private baseline is $48.1$. All our experiments suggest that larger models are better suited for private fine-tuning: while they are well known to achieve superior accuracy non-privately, we find that they also better maintain their accuracy when privacy is introduced.

Penulis (12)

D

Da Yu

S

Saurabh Naik

A

Arturs Backurs

S

Sivakanth Gopi

H

Huseyin A. Inan

G

Gautam Kamath

J

Janardhan Kulkarni

Y

Yin Tat Lee

A

Andre Manoel

L

Lukas Wutschitz

S

Sergey Yekhanin

H

Huishuai Zhang

Format Sitasi

Yu, D., Naik, S., Backurs, A., Gopi, S., Inan, H.A., Kamath, G. et al. (2021). Differentially Private Fine-tuning of Language Models. https://arxiv.org/abs/2110.06500

Akses Cepat

Lihat di Sumber
Informasi Jurnal
Tahun Terbit
2021
Bahasa
en
Sumber Database
arXiv
Akses
Open Access ✓