Quantifying Societal Bias Amplification in Image Captioning

Yusuke Hirota, Yuta Nakashima, Noa Garcia; Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2022, pp. 13450-13459

Abstract


We study societal bias amplification in image captioning. Image captioning models have been shown to perpetuate gender and racial biases, however, metrics to measure, quantify, and evaluate the societal bias in captions are not yet standardized. We provide a comprehensive study on the strengths and limitations of each metric, and propose LIC, a metric to study captioning bias amplification. We argue that, for image captioning, it is not enough to focus on the correct prediction of the protected attribute, and the whole context should be taken into account. We conduct extensive evaluation on traditional and state-of-the-art image captioning models, and surprisingly find that, by only focusing on the protected attribute prediction, bias mitigation models are unexpectedly amplifying bias.

Related Material


[pdf] [supp] [arXiv]
[bibtex]
@InProceedings{Hirota_2022_CVPR, author = {Hirota, Yusuke and Nakashima, Yuta and Garcia, Noa}, title = {Quantifying Societal Bias Amplification in Image Captioning}, booktitle = {Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)}, month = {June}, year = {2022}, pages = {13450-13459} }