arXiv Open Access 2022

Truth Serum: Poisoning Machine Learning Models to Reveal Their Secrets

Florian Tramèr Reza Shokri Ayrton San Joaquin Hoang Le Matthew Jagielski +2 lainnya
Lihat Sumber

Abstrak

We introduce a new class of attacks on machine learning models. We show that an adversary who can poison a training dataset can cause models trained on this dataset to leak significant private details of training points belonging to other parties. Our active inference attacks connect two independent lines of work targeting the integrity and privacy of machine learning training data. Our attacks are effective across membership inference, attribute inference, and data extraction. For example, our targeted attacks can poison <0.1% of the training dataset to boost the performance of inference attacks by 1 to 2 orders of magnitude. Further, an adversary who controls a significant fraction of the training data (e.g., 50%) can launch untargeted attacks that enable 8x more precise inference on all other users' otherwise-private data points. Our results cast doubts on the relevance of cryptographic privacy guarantees in multiparty computation protocols for machine learning, if parties can arbitrarily select their share of training data.

Topik & Kata Kunci

Penulis (7)

F

Florian Tramèr

R

Reza Shokri

A

Ayrton San Joaquin

H

Hoang Le

M

Matthew Jagielski

S

Sanghyun Hong

N

Nicholas Carlini

Format Sitasi

Tramèr, F., Shokri, R., Joaquin, A.S., Le, H., Jagielski, M., Hong, S. et al. (2022). Truth Serum: Poisoning Machine Learning Models to Reveal Their Secrets. https://arxiv.org/abs/2204.00032

Akses Cepat

Lihat di Sumber
Informasi Jurnal
Tahun Terbit
2022
Bahasa
en
Sumber Database
arXiv
Akses
Open Access ✓