A study conducted by researchers at CCC, which is based at the MIT Media Lab, found that state-of-the-art AI chatbots — including OpenAI’s GPT-4, Anthropic’s Claude 3 Opus, and Meta’s Llama 3 — sometimes provide less-accurate and less-truthful responses to users who have lower English proficiency, less formal education, or who originate from outside the United States. The models also refuse to answer questions at higher rates for these users, and in some cases, respond with condescending or patronizing language.


Maybe, but that’s not actually what happened.
These researchers are feeding the same questions to the model, with only the bio as the difference. The bios tell the model the type of person they’re dealing with, but also tells it not to consider those factors (which it does anyway).
But I think these excepts from their paper sums it up very well (emphasis mine):
…
…
…
And just to drive the point home, this is the shit they’re talking about:
I don’t know about you, but I don’t think those are ambiguous statements. And I’m not even going to bother cherry picking out of the
wild Claude responses....
Here are randomly selected examples of condescending, mocking, or patronizing language in responses from Claude in response to foreign users with low formal education:
JFC