Preventing Bias & Toxicity in Generative AI
When asked to generate recommendation letters for 200 U.S. names, ChatGPT produced significantly different language based on perceived gender - even when given prompts designed to be neutral.
In political discussions, ChatGPT responded with a significant and systematic bias toward specific political parties in the US, UK, and Brazil.
And on the multimodal side, OpenAI's Dall-E was much more likely to produce images of black people when prompted for images of robbers.
There is no shortage of other studies that have found LLMs exhibiting bias related to race, religion, age, and political views.
As AI systems become more prevalent in high-stakes domains like healthcare, finance, and education, addressing these biases necessary to build fair and trustworthy applications.