Is ChatGPT based?
Yes, it seems that ChatGPT does have biased responses regarding user's name, race, ethnic background, etc.
Personally, I don't think that varying responses based on the user's name and similar information is bad, as that's what we all do. If a man asks you for an advice, he should get a different response than the one a woman would get.
OpenAI claims that the bias takes less than 1% of total ChatGPT's responses and that very few of those are actually harmful stereotypes.
OpenAI's goal is to have a "fair" language model with as few biases, however, I think that a biased response is better tailored to the user needs and queries as long as the response doesn't include toxic and negative stereotypes.
Language Model Research Assistant - LMRA
To conduct this research, OpenAI is using a new "Language Model Research Assistant (LMRA)" that uses human feedback regarding ChatGPT's bias to diagnose further prompts and give feedback about the bias.