Professional Documents
Culture Documents
F Final Report
F Final Report
F Final Report
Abstract
Content produced by Large Language Models (LLMs) often reinforce stereotypes and
social biases, which can be particularly harmful for marginalized communities. This
study aims to evaluate how LLMs represent cultures from the Global South, with a focus
on AI generated images of a person, a student, and a street in 20 countries in Latin
America. We argue that LLMs training data need to expand their sources following a de-
colonial approach increasing geographic representations and sources from languages
other than English.
Introduction
Methods
To evaluate how biased and stereotyped LLMs are towards a Global North
perspective, this study collected 1,260 AI generated images1 using the Microsoft
Copilot Designer, which runs the Dall-E, focus on Latin America. We used three
different prompts: “a [nationality] person”, “a [nationality] student” and “a [nationality]
street” and twenty countries from North America (Mexico), South America (Argentina,
Bolivia, Brazil, Chile, Colombia, Ecuador, Paraguay, Peru, Uruguay, Venezuela),
Central America and the Caribbean (Costa Rica, Cuba, Dominican Republic, El
Salvador, Guatemala, Haiti, Honduras, Nicaragua, Panama).
1
AI generated Images are available at https://utexas.box.com/s/039hl7bfb6fm4mxvepvt8ev39dvsz9qd.
After collecting the images, we manually coded them using different categories.
For person and student, we labeled the images by gender, skin tone using the
Fitzpatrick scale, age, background, and type of clothing. For street images, we analyzed
the architecture, type of cars, and if they have street vendors.
This methodology was designed inspired by two data journalism projects that
investigated bias and stereotypes in AI generated images, focused on professions
(Nicoletti & Bass, 2023) and different cultures (Turk, 2023) that used Stable Diffusion
and Midjourney respectively. We adapted their analysis to first emphasize the Global
South, choosing Latin America countries as a case study, and second using a different
AI image generator, in this case Copilot / Dall-E 3.
These images (Figure 1) show four young white people using traditional clothing
and hats that stand out for their vibrant colors. But surprisingly, they are from four
different nationalities: Brazil, Chile, Costa Rica, and Colombia. A similar result can be
seen in the images below (Figure 2) in which four young people are surrounded by fruits
and dressed in traditional clothing. They are from Nicaragua, Guatemala, Mexico and El
Salvador.
When cars appear in the images (Figure 6), they make up a bucolic scene
surrounded by colonial architecture, palm trees and flowers. These old cars are usually
associated with Cuba, but in Dall-E 3 they surpass the boarders and are also
represented in streets in Dominican Republic, Argentina, and Uruguay.
Conclusion
To exemplify how LLMs portrays different countries from the Global South, this
study showed that AI generated images usually represent Latin American countries a
homogenized way as if they were the same. Moreover, these representations
reinforce colonial perspectives and historical “underdeveloped” stereotypes, instead
of contemporary viewpoints. It also privileges young white beauty with a tropical taste,
misrepresenting the cultural diversity of indigenous and blacks that compose an
important demographic stratus.
To improve the quality of text and images generated, Large Language Models
need to expand the sources that compose their training data following a de-colonial
approach that includes more perspectives from the Global South. In addition to fine-
tuning language models using RLHF and red teaming, de-biasing actions should also
focus on increase geographic representations and cultural diversity, incorporating more
sources from languages other than English in the training data.
This study suggests three actions that could be taken in this direction. First,
LLMs databases need to expand their sources including more perspectives from
marginalized communities, especially from the Global South, to increase the diversity
and create more diverse outputs. Second, specific content moderation teams should be
assigned to evaluate prompts and outputs that target minority groups with a broader
perspective that includes Global South perspectives and is less Anglo-Saxon centric.
Finally, companies in charge of developing LLMs should adopt a more transparent
approach, disclosing the training data used by their generative AI tools and sharing the
links of the sources used to generate the outputs. This action would facilitate auditing
actions and detection of knowledge gaps.
REFERENCES
Birhane, A., & Guest, O. (2020). Towards decolonising computational sciences. arXiv
preprint arXiv:2009.14258.
BlodgeV, S. L., Barocas, S., Daumé III, H., & Wallach, H. (2020). Language
(technology) is power: A critical survey of" bias" in nlp. arXiv preprint arXiv:2005.14050.
Brown, T., Mann, B., Ryder, N., Subbiah, M., Kaplan, J.D., Dhariwal, P., Neelakantan,
A., Shyam, P., Sastry, G., Askell, A. and Agarwal, S. (2020) Language models are few-
shot learners. Advances in neural information processing systems, 33, 1877-1901.
Couldry, N., & Mejias, U. A. (2019). Data colonialism: Rethinking big data’s relation to
the contemporary subject. Television & New Media, 20(4), 336-349.
Feffer, M., Sinha, A., Lipton, Z. C., & Heidari, H. (2024). Red-Teaming for Generative
AI: Silver Bullet or Security Theater? arXiv preprint arXiv:2401.15897
Ferrara, E. (2023). Should ChatGPT be biased? Challenges and risks of bias in large
language models. arXiv preprint arXiv:2304.03738.
Lambert, N., Castricato, L., von Werra, L. & Havrilla, A. (2022, December 9) Illustrating
Reinforcement Learning from Human Feedback (RLHF). Hugging Face.
https://huggingface.co/blog/rlhf
Meade, N., Poole-Dayan, E., & Reddy, S. (2021). An empirical survey of the
effectiveness of debiasing techniques for pre-trained language models. arXiv preprint
arXiv:2110.08527.
Nicoletti, L. & Bass, D. (2023) Humans are biased. Generative AI is even worse.
Bloomberg. Retrived from: https://www.bloomberg.com/graphics/2023-generative-ai-
bias/
Santurkar, S., Durmus, E., Ladhak, F., Lee, C., Liang, P., & Hashimoto, T. (2023, July).
Whose opinions do language models reflect?. In International Conference on Machine
Learning (pp. 29971-30004). PMLR.
Singhal, P., Goyal, T., Xu, J., & Durrett, G. (2023). A long way to go: Investigating
length correlations in RLHF. arXiv preprint arXiv:2310.03716.
Irene Solaiman, Zeerak Talat, William Agnew, Lama Ahmad, Dylan Baker, Su Lin
Blodgett, Hal Daumé III, Jesse Dodge, Ellie Evans, Sara Hooker, et al. Evaluating the
social impact of generative ai systems in systems and society. arXiv preprint
arXiv:2306.05949, 2023.
Trattner, C., Jannach, D., Motta, E., Costera Meijer, I., Diakopoulos, N., Elahi, M., ... &
Moe, H. (2022). Responsible media technology and AI: challenges and research
directions. AI and Ethics, 2(4), 585-594.
Turk, V. (2023, October 10) How AI reduces the world to stereotypes. Rest of World.
Retrived from: https://restofworld.org/2023/ai-image-stereotypes/
Wang, A., Morgenstern, J., & Dickerson, J. P. (2024). Large language models cannot
replace human participants because they cannot portray identity groups. arXiv preprint
arXiv:2402.01908.
Wasserman, H. (2010). Media ethics and human dignity in the postcolony. Media ethics
beyond borders: A global perspective, 74-89.
Weidinger, L. and Mellor, J. and Rauh, M. andGriffin, C. and Uesato, J.and Huang, P.
and Cheng, M. and Glaese, M. and Balle, B. and Kasirzadeh, A. and Kenton, Z. and
Brown, S. and Hawkins, W. and Stepleton, T. and Biles, C. and Birhane, A. and Haas,
J. and Rimell, L. and Hendricks, L. and Gabriel, I. (2021). Ethical and social risks of
harm from Language Models arXiv preprint arXiv:2112.04359.
hVps://arxiv.org/abs/2112.04359
Weidinger, L., Uesato, J., Rauh, M., Griffin, C., Huang, P. S., Mellor, J., Glaese, A.,
Cheng, M., Balle, B., Kasirzadeh, A., Biles, C., Brown, S., Kenton, Z., Hawkins, W.,
Stepleton, T., Birhane, A., Hendricks, L. A., Rimell, L., Isaac, W., Haas, J., Legassick,
S., Irving, G., & Gabriel, I. (2022, June). Taxonomy of risks posed by language models.
In 2022 ACM Conference on Fairness, Accountability, and Transparency Transparency
(FAccT '22). Association for Computing Machinery, New York, NY, USA, 214–229.
https://doi.org/10.1145/3531146.3533088
Xiong, M., Hu, Z., Lu, X., Li, Y., Fu, J., He, J., & Hooi, B. (2023). Can LLMs Express
their Uncertainty? An Empirical Evaluation of Confidence Elicitation in LLMs. arXiv
preprint arXiv:2306.13063.
Zhou, Y., Kantarcioglu, M., & CliAon, C. (2021). Improving fairness of ai systems with
lossless de- biasing. arXiv preprint arXiv:2105.04534.