Subscribe now

Technology

OpenAI’s chatbot shows racial bias in advising home buyers and renters

ChatGPT often suggests lower-income neighbourhoods to people who are Black, showing prejudices reflecting generations of housing discrimination in the US

By Jeremy Hsu

21 May 2024

B8EP7D Aerial view of houses located in Staten Island, New York., USA

AI models could reinforce housing discrimination in relatively segregated cities like New York

Robert Quinlan/Alamy

An AI chatbot often recommends neighbourhoods to potential home buyers and renters based on race. This may be a reflection of bias stemming from generations of US housing discrimination.

“A lot of people think that generative AI and large language models are the emerging technologies of the future,” says Eric Liu at the Massachusetts Institute of Technology. “But of course they’re being trained on data from the past.”

Liu and…

Sign up to our weekly newsletter

Receive a weekly dose of discovery in your inbox. We'll also keep you up to date with New Scientist events and special offers.

Sign up

To continue reading, subscribe today with our introductory offers

Popular articles

Trending New Scientist articles

Piano Exit Overlay Banner Mobile Piano Exit Overlay Banner Desktop