Exploring LGBTQ+ Bias in Generative AI Answers across Different Country and Religious Contexts

L Vicsek, A Vancsó, M Zajko, J Takacs - arXiv preprint arXiv:2407.03473, 2024 - arxiv.org
L Vicsek, A Vancsó, M Zajko, J Takacs
arXiv preprint arXiv:2407.03473, 2024arxiv.org
Previous discussions have highlighted the need for generative AI tools to become more
culturally sensitive, yet often neglect the complexities of handling content about minorities,
who are perceived differently across cultures and religions. Our study examined how two
generative AI systems respond to homophobic statements with varying cultural and religious
context information. Findings showed ChatGPT 3.5's replies exhibited cultural relativism, in
contrast to Bard's, which stressed human rights and provided more support for LGBTQ+ …
Previous discussions have highlighted the need for generative AI tools to become more culturally sensitive, yet often neglect the complexities of handling content about minorities, who are perceived differently across cultures and religions. Our study examined how two generative AI systems respond to homophobic statements with varying cultural and religious context information. Findings showed ChatGPT 3.5's replies exhibited cultural relativism, in contrast to Bard's, which stressed human rights and provided more support for LGBTQ+ issues. Both demonstrated significant change in responses based on contextual information provided in the prompts, suggesting that AI systems may adjust in their responses the degree and forms of support for LGBTQ+ people according to information they receive about the user's background. The study contributes to understanding the social and ethical implications of AI responses and argues that any work to make generative AI outputs more culturally diverse requires a grounding in fundamental human rights.
arxiv.org