Semantic Scholar Open Access 2020 1699 sitasi

Gradient Surgery for Multi-Task Learning

Tianhe Yu Saurabh Kumar Abhishek Gupta S. Levine Karol Hausman +1 lainnya

Abstrak

While deep learning and deep reinforcement learning (RL) systems have demonstrated impressive results in domains such as image classification, game playing, and robotic control, data efficiency remains a major challenge. Multi-task learning has emerged as a promising approach for sharing structure across multiple tasks to enable more efficient learning. However, the multi-task setting presents a number of optimization challenges, making it difficult to realize large efficiency gains compared to learning tasks independently. The reasons why multi-task learning is so challenging compared to single-task learning are not fully understood. In this work, we identify a set of three conditions of the multi-task optimization landscape that cause detrimental gradient interference, and develop a simple yet general approach for avoiding such interference between task gradients. We propose a form of gradient surgery that projects a task's gradient onto the normal plane of the gradient of any other task that has a conflicting gradient. On a series of challenging multi-task supervised and multi-task RL problems, this approach leads to substantial gains in efficiency and performance. Further, it is model-agnostic and can be combined with previously-proposed multi-task architectures for enhanced performance.

Penulis (6)

T

Tianhe Yu

S

Saurabh Kumar

A

Abhishek Gupta

S

S. Levine

K

Karol Hausman

C

Chelsea Finn

Format Sitasi

Yu, T., Kumar, S., Gupta, A., Levine, S., Hausman, K., Finn, C. (2020). Gradient Surgery for Multi-Task Learning. https://www.semanticscholar.org/paper/449c5660d637741f7aa7ff42549c32b43c9968bf

Akses Cepat

PDF tidak tersedia langsung

Cek di sumber asli →
Lihat di Sumber
Informasi Jurnal
Tahun Terbit
2020
Bahasa
en
Total Sitasi
1699×
Sumber Database
Semantic Scholar
Akses
Open Access ✓