BBVA’s Stress Test Highlights Bias in Generative AI Models

January 3, 2025

In today’s technologically advanced world, the capabilities of generative AI models like GPT, Gemini, and Llama are transforming various sectors by enabling users to directly interact with artificial intelligence. However, the same potential that empowers users also poses significant challenges, as the responses generated by these models can reveal considerable biases. Such biases are tied to a range of factors, including gender, race, sexual orientation, and disability, which are often a byproduct of the data and training processes used. Although developers work diligently to curb these biases, the predominance of English in training datasets makes it challenging to achieve bias reduction effectively across different languages. This issue underscores the need for comprehensive measures to ensure that AI models do not reinforce harmful societal stereotypes.

The Need to Address Biased Algorithms

The integration of AI into critical areas like workplaces, justice systems, and healthcare necessitates an urgent focus on unbiased algorithms, as decisions taken in these domains significantly influence individuals’ lives and societal structures. AI’s increasing role in decision-making accentuates the risk of propagating biases if not properly addressed. These biases often arise from distortions in the training data, reflecting existing prejudices and potentially leading to discriminatory outcomes. To counteract these effects, proposals have been presented to caution against bias in AI applications, advocating for rigorous testing and evaluation methods to minimize these risks.

In response to the challenge of biased AI responses in languages other than English, IBM Research developed a specialized dataset called SocialStigmaQA (SSQA). This dataset aims to measure how generative models express biases connected to various stigmas. BBVA has adapted this dataset to Spanish, while IBM extended it to Japanese, providing a broader view of AI biases across different languages. Initial tests revealed that biased responses were more prevalent in Spanish and Japanese compared to their English counterparts. The dataset encompasses around one hundred stigma conditions, combined with forty hypothetical scenarios, creating user prompts to evaluate AI responses. By benchmarking these responses against predefined biased answers, researchers can gauge the extent of discrimination exhibited by generative models.

Implementation and Importance of Stress Tests

The SSQA dataset serves as a ‘stress test’ to rigorously challenge AI models, bringing to light the biases they might harbor. This approach plays a crucial role in fostering balanced generative AI systems, ensuring that they reflect the diverse cultural and social realities of various linguistic regions. Clara Higuera, a prominent study author and data scientist at BBVA’s GenAI Lab, emphasizes the significance of such analysis in promoting secure and responsible use of generative AI. She accentuates the need for further research to understand and mitigate biases comprehensively.

The importance of this work was highlighted during its presentation at the NeurIPS conference, where BBVA and IBM’s efforts were recognized for their commitment to socially responsible language modeling. By releasing the Spanish and Japanese datasets on open-source platforms like GitHub and HuggingFace, the researchers have invited the global community to contribute to improving these resources. This collaborative effort ensures ongoing enhancement and adaptation of the datasets to include additional stigmas from diverse sources, such as the European Social Survey. The goal is to continuously evolve the datasets to better address biases and reflect more comprehensive societal contexts.

The Path Forward for Fair Generative AI

The SSQA dataset acts as a rigorous test to challenge AI models, revealing potential biases. This effort is crucial for developing balanced generative AI systems that reflect the varied cultural and social realities of different linguistic regions. Clara Higuera, a leading author and data scientist at BBVA’s GenAI Lab, highlights the importance of such analysis for the safe and responsible use of generative AI. She emphasizes the necessity for ongoing research to thoroughly understand and mitigate biases.

The significance of this work was underscored at the NeurIPS conference, where BBVA and IBM were recognized for their dedication to socially responsible language modeling. By making the Spanish and Japanese datasets available on open-source platforms like GitHub and HuggingFace, researchers have encouraged the global community to contribute to enhancing these resources. This collaborative approach ensures that the datasets evolve to include biases identified by diverse sources such as the European Social Survey. The aim is for the datasets to continually develop, improving their ability to address biases and better reflect comprehensive societal contexts.

Subscribe to our weekly news digest.

Join now and become a part of our fast-growing community.

Invalid Email Address
Thanks for Subscribing!
We'll be sending you our best soon!
Something went wrong, please try again later