Not the usual tropes again...
If you look for racism and sexism, you'll find it. Even if it's not there.
What the training set has done is find the general composition of the set, and render that. If that's what the contribution to the set is, that's what'll be reflected.
I'm always confounded by AI workers coming up with these exclamations after exposing learning mechanisms to a representative set of data and it doesn't come up with what they want. If you want to teach something to come up with answers you want, you need to put the effort into creating a curated set of information for it to learn from (this is something that every species on the planet has learned long ago, which is why they survive and in cases such as some Birds and Simians, develop their own actual cultures.
But, when you select only what you want to see, then you have to understand that it's a product of your own biases. Once you take off the limiters and let it see what's out there, it'll learn things you'd rather it didn't.
There are, of course, confounding issues (like contrast for some human gene adaptations, which make some faces harder to recognise for example), but those are technical hurdles which you can control for and work with (it's part of learning how to teach a learning system).