Semantic Scholar Open Access 2025 2 sitasi

Uncovering Bias in Foundation Models: Impact, Testing, Harm, and Mitigation

Shuzhou Sun Li Liu Yongxiang Liu Zhen Liu Shuanghui Zhang +11 lainnya

Abstrak

Bias in Foundation Models (FMs) - trained on vast datasets spanning societal and historical knowledge - poses significant challenges for fairness and equity across fields such as healthcare, education, and finance. These biases, rooted in the overrepresentation of stereotypes and societal inequalities in training data, exacerbate real-world discrimination, reinforce harmful stereotypes, and erode trust in AI systems. To address this, we introduce Trident Probe Testing (TriProTesting), a systematic testing method that detects explicit and implicit biases using semantically designed probes. Here we show that FMs, including CLIP, ALIGN, BridgeTower, and OWLv2, demonstrate pervasive biases across single and mixed social attributes (gender, race, age, and occupation). Notably, we uncover mixed biases when social attributes are combined, such as gender x race, gender x age, and gender x occupation, revealing deeper layers of discrimination. We further propose Adaptive Logit Adjustment (AdaLogAdjustment), a post-processing technique that dynamically redistributes probability power to mitigate these biases effectively, achieving significant improvements in fairness without retraining models. These findings highlight the urgent need for ethical AI practices and interdisciplinary solutions to address biases not only at the model level but also in societal structures. Our work provides a scalable and interpretable solution that advances fairness in AI systems while offering practical insights for future research on fair AI technologies.

Topik & Kata Kunci

Penulis (16)

S

Shuzhou Sun

L

Li Liu

Y

Yongxiang Liu

Z

Zhen Liu

S

Shuanghui Zhang

J

J. Heikkila

X

Xiang Li The College of Computer Science

N

Nankai University

T

Tianjin

C

China

T

The Center for Machine Vision

S

Signal Analysis

U

U. Oulu

F

Finland.

T

T. Z. F. O. Science

N

National Institute of Technology

Format Sitasi

Sun, S., Liu, L., Liu, Y., Liu, Z., Zhang, S., Heikkila, J. et al. (2025). Uncovering Bias in Foundation Models: Impact, Testing, Harm, and Mitigation. https://www.semanticscholar.org/paper/4bbdbb9268247082d5b4208fe5270dbdb172eb5e

Akses Cepat

PDF tidak tersedia langsung

Cek di sumber asli →
Lihat di Sumber
Informasi Jurnal
Tahun Terbit
2025
Bahasa
en
Total Sitasi
Sumber Database
Semantic Scholar
Akses
Open Access ✓