arXiv Open Access 2024

AI red-teaming is a sociotechnical problem: on values, labor, and harms

Tarleton Gillespie Ryland Shaw Mary L. Gray Jina Suh
Lihat Sumber

Abstrak

As generative AI technologies find more and more real-world applications, the importance of testing their performance and safety seems paramount. "Red-teaming" has quickly become the primary approach to test AI models--prioritized by AI companies, and enshrined in AI policy and regulation. Members of red teams act as adversaries, probing AI systems to test their safety mechanisms and uncover vulnerabilities. Yet we know far too little about this work or its implications. This essay calls for collaboration between computer scientists and social scientists to study the sociotechnical systems surrounding AI technologies, including the work of red-teaming, to avoid repeating the mistakes of the recent past. We highlight the importance of understanding the values and assumptions behind red-teaming, the labor arrangements involved, and the psychological impacts on red-teamers, drawing insights from the lessons learned around the work of content moderation.

Topik & Kata Kunci

Penulis (4)

T

Tarleton Gillespie

R

Ryland Shaw

M

Mary L. Gray

J

Jina Suh

Format Sitasi

Gillespie, T., Shaw, R., Gray, M.L., Suh, J. (2024). AI red-teaming is a sociotechnical problem: on values, labor, and harms. https://arxiv.org/abs/2412.09751

Akses Cepat

Lihat di Sumber
Informasi Jurnal
Tahun Terbit
2024
Bahasa
en
Sumber Database
arXiv
Akses
Open Access ✓